GithubHelp home page GithubHelp logo

seaweedfs / seaweedfs Goto Github PK

View Code? Open in Web Editor NEW
21.1K 537.0 2.2K 54.92 MB

SeaweedFS is a fast distributed storage system for blobs, objects, files, and data lake, for billions of files! Blob store has O(1) disk seek, cloud tiering. Filer supports Cloud Drive, cross-DC active-active replication, Kubernetes, POSIX FUSE mount, S3 API, S3 Gateway, Hadoop, WebDAV, encryption, Erasure Coding.

License: Apache License 2.0

Go 88.27% Makefile 0.26% Shell 0.17% Java 10.39% Smarty 0.17% HTML 0.68% Dockerfile 0.02% Lua 0.04%
distributed-storage distributed-systems s3 hdfs fuse distributed-file-system hadoop-hdfs posix tiered-file-system kubernetes

seaweedfs's Introduction

SeaweedFS

Slack Twitter Build Status GoDoc Wiki Docker Pulls SeaweedFS on Maven Central Artifact Hub

SeaweedFS Logo

SeaweedFS is an independent Apache-licensed open source project with its ongoing development made possible entirely thanks to the support of these awesome backers. If you'd like to grow SeaweedFS even stronger, please consider joining our sponsors on Patreon.

Your support will be really appreciated by me and other supporters!

Gold Sponsors

nodion piknik


Table of Contents

Quick Start

Quick Start for S3 API on Docker

docker run -p 8333:8333 chrislusf/seaweedfs server -s3

Quick Start with Single Binary

  • Download the latest binary from https://github.com/seaweedfs/seaweedfs/releases and unzip a single binary file weed or weed.exe. Or run go install github.com/seaweedfs/seaweedfs/weed@latest.
  • Run weed server -dir=/some/data/dir -s3 to start one master, one volume server, one filer, and one S3 gateway.

Also, to increase capacity, just add more volume servers by running weed volume -dir="/some/data/dir2" -mserver="<master_host>:9333" -port=8081 locally, or on a different machine, or on thousands of machines. That is it!

Quick Start SeaweedFS S3 on AWS

Introduction

SeaweedFS is a simple and highly scalable distributed file system. There are two objectives:

  1. to store billions of files!
  2. to serve the files fast!

SeaweedFS started as an Object Store to handle small files efficiently. Instead of managing all file metadata in a central master, the central master only manages volumes on volume servers, and these volume servers manage files and their metadata. This relieves concurrency pressure from the central master and spreads file metadata into volume servers, allowing faster file access (O(1), usually just one disk read operation).

There is only 40 bytes of disk storage overhead for each file's metadata. It is so simple with O(1) disk reads that you are welcome to challenge the performance with your actual use cases.

SeaweedFS started by implementing Facebook's Haystack design paper. Also, SeaweedFS implements erasure coding with ideas from f4: Facebook’s Warm BLOB Storage System, and has a lot of similarities with Facebook’s Tectonic Filesystem

On top of the object store, optional Filer can support directories and POSIX attributes. Filer is a separate linearly-scalable stateless server with customizable metadata stores, e.g., MySql, Postgres, Redis, Cassandra, HBase, Mongodb, Elastic Search, LevelDB, RocksDB, Sqlite, MemSql, TiDB, Etcd, CockroachDB, YDB, etc.

For any distributed key value stores, the large values can be offloaded to SeaweedFS. With the fast access speed and linearly scalable capacity, SeaweedFS can work as a distributed Key-Large-Value store.

SeaweedFS can transparently integrate with the cloud. With hot data on local cluster, and warm data on the cloud with O(1) access time, SeaweedFS can achieve both fast local access time and elastic cloud storage capacity. What's more, the cloud storage access API cost is minimized. Faster and cheaper than direct cloud storage!

Back to TOC

Features

Additional Features

  • Can choose no replication or different replication levels, rack and data center aware.
  • Automatic master servers failover - no single point of failure (SPOF).
  • Automatic Gzip compression depending on file MIME type.
  • Automatic compaction to reclaim disk space after deletion or update.
  • Automatic entry TTL expiration.
  • Any server with some disk space can add to the total storage space.
  • Adding/Removing servers does not cause any data re-balancing unless triggered by admin commands.
  • Optional picture resizing.
  • Support ETag, Accept-Range, Last-Modified, etc.
  • Support in-memory/leveldb/readonly mode tuning for memory/performance balance.
  • Support rebalancing the writable and readonly volumes.
  • Customizable Multiple Storage Tiers: Customizable storage disk types to balance performance and cost.
  • Transparent cloud integration: unlimited capacity via tiered cloud storage for warm data.
  • Erasure Coding for warm storage Rack-Aware 10.4 erasure coding reduces storage cost and increases availability.

Back to TOC

Filer Features

Kubernetes

Back to TOC

Example: Using Seaweed Object Store

By default, the master node runs on port 9333, and the volume nodes run on port 8080. Let's start one master node, and two volume nodes on port 8080 and 8081. Ideally, they should be started from different machines. We'll use localhost as an example.

SeaweedFS uses HTTP REST operations to read, write, and delete. The responses are in JSON or JSONP format.

Start Master Server

> ./weed master

Start Volume Servers

> weed volume -dir="/tmp/data1" -max=5  -mserver="localhost:9333" -port=8080 &
> weed volume -dir="/tmp/data2" -max=10 -mserver="localhost:9333" -port=8081 &

Write File

To upload a file: first, send a HTTP POST, PUT, or GET request to /dir/assign to get an fid and a volume server URL:

> curl http://localhost:9333/dir/assign
{"count":1,"fid":"3,01637037d6","url":"127.0.0.1:8080","publicUrl":"localhost:8080"}

Second, to store the file content, send a HTTP multi-part POST request to url + '/' + fid from the response:

> curl -F file=@/home/chris/myphoto.jpg http://127.0.0.1:8080/3,01637037d6
{"name":"myphoto.jpg","size":43234,"eTag":"1cc0118e"}

To update, send another POST request with updated file content.

For deletion, send an HTTP DELETE request to the same url + '/' + fid URL:

> curl -X DELETE http://127.0.0.1:8080/3,01637037d6

Save File Id

Now, you can save the fid, 3,01637037d6 in this case, to a database field.

The number 3 at the start represents a volume id. After the comma, it's one file key, 01, and a file cookie, 637037d6.

The volume id is an unsigned 32-bit integer. The file key is an unsigned 64-bit integer. The file cookie is an unsigned 32-bit integer, used to prevent URL guessing.

The file key and file cookie are both coded in hex. You can store the <volume id, file key, file cookie> tuple in your own format, or simply store the fid as a string.

If stored as a string, in theory, you would need 8+1+16+8=33 bytes. A char(33) would be enough, if not more than enough, since most uses will not need 2^32 volumes.

If space is really a concern, you can store the file id in your own format. You would need one 4-byte integer for volume id, 8-byte long number for file key, and a 4-byte integer for the file cookie. So 16 bytes are more than enough.

Read File

Here is an example of how to render the URL.

First look up the volume server's URLs by the file's volumeId:

> curl http://localhost:9333/dir/lookup?volumeId=3
{"volumeId":"3","locations":[{"publicUrl":"localhost:8080","url":"localhost:8080"}]}

Since (usually) there are not too many volume servers, and volumes don't move often, you can cache the results most of the time. Depending on the replication type, one volume can have multiple replica locations. Just randomly pick one location to read.

Now you can take the public URL, render the URL or directly read from the volume server via URL:

 http://localhost:8080/3,01637037d6.jpg

Notice we add a file extension ".jpg" here. It's optional and just one way for the client to specify the file content type.

If you want a nicer URL, you can use one of these alternative URL formats:

 http://localhost:8080/3/01637037d6/my_preferred_name.jpg
 http://localhost:8080/3/01637037d6.jpg
 http://localhost:8080/3,01637037d6.jpg
 http://localhost:8080/3/01637037d6
 http://localhost:8080/3,01637037d6

If you want to get a scaled version of an image, you can add some params:

http://localhost:8080/3/01637037d6.jpg?height=200&width=200
http://localhost:8080/3/01637037d6.jpg?height=200&width=200&mode=fit
http://localhost:8080/3/01637037d6.jpg?height=200&width=200&mode=fill

Rack-Aware and Data Center-Aware Replication

SeaweedFS applies the replication strategy at a volume level. So, when you are getting a file id, you can specify the replication strategy. For example:

curl http://localhost:9333/dir/assign?replication=001

The replication parameter options are:

000: no replication
001: replicate once on the same rack
010: replicate once on a different rack, but same data center
100: replicate once on a different data center
200: replicate twice on two different data center
110: replicate once on a different rack, and once on a different data center

More details about replication can be found on the wiki.

You can also set the default replication strategy when starting the master server.

Allocate File Key on Specific Data Center

Volume servers can be started with a specific data center name:

 weed volume -dir=/tmp/1 -port=8080 -dataCenter=dc1
 weed volume -dir=/tmp/2 -port=8081 -dataCenter=dc2

When requesting a file key, an optional "dataCenter" parameter can limit the assigned volume to the specific data center. For example, this specifies that the assigned volume should be limited to 'dc1':

 http://localhost:9333/dir/assign?dataCenter=dc1

Other Features

Back to TOC

Object Store Architecture

Usually distributed file systems split each file into chunks, a central master keeps a mapping of filenames, chunk indices to chunk handles, and also which chunks each chunk server has.

The main drawback is that the central master can't handle many small files efficiently, and since all read requests need to go through the chunk master, so it might not scale well for many concurrent users.

Instead of managing chunks, SeaweedFS manages data volumes in the master server. Each data volume is 32GB in size, and can hold a lot of files. And each storage node can have many data volumes. So the master node only needs to store the metadata about the volumes, which is a fairly small amount of data and is generally stable.

The actual file metadata is stored in each volume on volume servers. Since each volume server only manages metadata of files on its own disk, with only 16 bytes for each file, all file access can read file metadata just from memory and only needs one disk operation to actually read file data.

For comparison, consider that an xfs inode structure in Linux is 536 bytes.

Master Server and Volume Server

The architecture is fairly simple. The actual data is stored in volumes on storage nodes. One volume server can have multiple volumes, and can both support read and write access with basic authentication.

All volumes are managed by a master server. The master server contains the volume id to volume server mapping. This is fairly static information, and can be easily cached.

On each write request, the master server also generates a file key, which is a growing 64-bit unsigned integer. Since write requests are not generally as frequent as read requests, one master server should be able to handle the concurrency well.

Write and Read files

When a client sends a write request, the master server returns (volume id, file key, file cookie, volume node URL) for the file. The client then contacts the volume node and POSTs the file content.

When a client needs to read a file based on (volume id, file key, file cookie), it asks the master server by the volume id for the (volume node URL, volume node public URL), or retrieves this from a cache. Then the client can GET the content, or just render the URL on web pages and let browsers fetch the content.

Please see the example for details on the write-read process.

Storage Size

In the current implementation, each volume can hold 32 gibibytes (32GiB or 8x2^32 bytes). This is because we align content to 8 bytes. We can easily increase this to 64GiB, or 128GiB, or more, by changing 2 lines of code, at the cost of some wasted padding space due to alignment.

There can be 4 gibibytes (4GiB or 2^32 bytes) of volumes. So the total system size is 8 x 4GiB x 4GiB which is 128 exbibytes (128EiB or 2^67 bytes).

Each individual file size is limited to the volume size.

Saving memory

All file meta information stored on a volume server is readable from memory without disk access. Each file takes just a 16-byte map entry of <64bit key, 32bit offset, 32bit size>. Of course, each map entry has its own space cost for the map. But usually the disk space runs out before the memory does.

Tiered Storage to the cloud

The local volume servers are much faster, while cloud storages have elastic capacity and are actually more cost-efficient if not accessed often (usually free to upload, but relatively costly to access). With the append-only structure and O(1) access time, SeaweedFS can take advantage of both local and cloud storage by offloading the warm data to the cloud.

Usually hot data are fresh and warm data are old. SeaweedFS puts the newly created volumes on local servers, and optionally upload the older volumes on the cloud. If the older data are accessed less often, this literally gives you unlimited capacity with limited local servers, and still fast for new data.

With the O(1) access time, the network latency cost is kept at minimum.

If the hot/warm data is split as 20/80, with 20 servers, you can achieve storage capacity of 100 servers. That's a cost saving of 80%! Or you can repurpose the 80 servers to store new data also, and get 5X storage throughput.

Back to TOC

Compared to Other File Systems

Most other distributed file systems seem more complicated than necessary.

SeaweedFS is meant to be fast and simple, in both setup and operation. If you do not understand how it works when you reach here, we've failed! Please raise an issue with any questions or update this file with clarifications.

SeaweedFS is constantly moving forward. Same with other systems. These comparisons can be outdated quickly. Please help to keep them updated.

Back to TOC

Compared to HDFS

HDFS uses the chunk approach for each file, and is ideal for storing large files.

SeaweedFS is ideal for serving relatively smaller files quickly and concurrently.

SeaweedFS can also store extra large files by splitting them into manageable data chunks, and store the file ids of the data chunks into a meta chunk. This is managed by "weed upload/download" tool, and the weed master or volume servers are agnostic about it.

Back to TOC

Compared to GlusterFS, Ceph

The architectures are mostly the same. SeaweedFS aims to store and read files fast, with a simple and flat architecture. The main differences are

  • SeaweedFS optimizes for small files, ensuring O(1) disk seek operation, and can also handle large files.
  • SeaweedFS statically assigns a volume id for a file. Locating file content becomes just a lookup of the volume id, which can be easily cached.
  • SeaweedFS Filer metadata store can be any well-known and proven data store, e.g., Redis, Cassandra, HBase, Mongodb, Elastic Search, MySql, Postgres, Sqlite, MemSql, TiDB, CockroachDB, Etcd, YDB etc, and is easy to customize.
  • SeaweedFS Volume server also communicates directly with clients via HTTP, supporting range queries, direct uploads, etc.
System File Metadata File Content Read POSIX REST API Optimized for large number of small files
SeaweedFS lookup volume id, cacheable O(1) disk seek Yes Yes
SeaweedFS Filer Linearly Scalable, Customizable O(1) disk seek FUSE Yes Yes
GlusterFS hashing FUSE, NFS
Ceph hashing + rules FUSE Yes
MooseFS in memory FUSE No
MinIO separate meta file for each file Yes No

Back to TOC

Compared to GlusterFS

GlusterFS stores files, both directories and content, in configurable volumes called "bricks".

GlusterFS hashes the path and filename into ids, and assigned to virtual volumes, and then mapped to "bricks".

Back to TOC

Compared to MooseFS

MooseFS chooses to neglect small file issue. From moosefs 3.0 manual, "even a small file will occupy 64KiB plus additionally 4KiB of checksums and 1KiB for the header", because it "was initially designed for keeping large amounts (like several thousands) of very big files"

MooseFS Master Server keeps all meta data in memory. Same issue as HDFS namenode.

Back to TOC

Compared to Ceph

Ceph can be setup similar to SeaweedFS as a key->blob store. It is much more complicated, with the need to support layers on top of it. Here is a more detailed comparison

SeaweedFS has a centralized master group to look up free volumes, while Ceph uses hashing and metadata servers to locate its objects. Having a centralized master makes it easy to code and manage.

Ceph, like SeaweedFS, is based on the object store RADOS. Ceph is rather complicated with mixed reviews.

Ceph uses CRUSH hashing to automatically manage data placement, which is efficient to locate the data. But the data has to be placed according to the CRUSH algorithm. Any wrong configuration would cause data loss. Topology changes, such as adding new servers to increase capacity, will cause data migration with high IO cost to fit the CRUSH algorithm. SeaweedFS places data by assigning them to any writable volumes. If writes to one volume failed, just pick another volume to write. Adding more volumes is also as simple as it can be.

SeaweedFS is optimized for small files. Small files are stored as one continuous block of content, with at most 8 unused bytes between files. Small file access is O(1) disk read.

SeaweedFS Filer uses off-the-shelf stores, such as MySql, Postgres, Sqlite, Mongodb, Redis, Elastic Search, Cassandra, HBase, MemSql, TiDB, CockroachCB, Etcd, YDB, to manage file directories. These stores are proven, scalable, and easier to manage.

SeaweedFS comparable to Ceph advantage
Master MDS simpler
Volume OSD optimized for small files
Filer Ceph FS linearly scalable, Customizable, O(1) or O(logN)

Back to TOC

Compared to MinIO

MinIO follows AWS S3 closely and is ideal for testing for S3 API. It has good UI, policies, versionings, etc. SeaweedFS is trying to catch up here. It is also possible to put MinIO as a gateway in front of SeaweedFS later.

MinIO metadata are in simple files. Each file write will incur extra writes to corresponding meta file.

MinIO does not have optimization for lots of small files. The files are simply stored as is to local disks. Plus the extra meta file and shards for erasure coding, it only amplifies the LOSF problem.

MinIO has multiple disk IO to read one file. SeaweedFS has O(1) disk reads, even for erasure coded files.

MinIO has full-time erasure coding. SeaweedFS uses replication on hot data for faster speed and optionally applies erasure coding on warm data.

MinIO does not have POSIX-like API support.

MinIO has specific requirements on storage layout. It is not flexible to adjust capacity. In SeaweedFS, just start one volume server pointing to the master. That's all.

Dev Plan

  • More tools and documentation, on how to manage and scale the system.
  • Read and write stream data.
  • Support structured data.

This is a super exciting project! And we need helpers and support!

Back to TOC

Installation Guide

Installation guide for users who are not familiar with golang

Step 1: install go on your machine and setup the environment by following the instructions at:

https://golang.org/doc/install

make sure to define your $GOPATH

Step 2: checkout this repo:

git clone https://github.com/seaweedfs/seaweedfs.git

Step 3: download, compile, and install the project by executing the following command

cd seaweedfs/weed && make install

Once this is done, you will find the executable "weed" in your $GOPATH/bin directory

Back to TOC

Disk Related Topics

Hard Drive Performance

When testing read performance on SeaweedFS, it basically becomes a performance test of your hard drive's random read speed. Hard drives usually get 100MB/s~200MB/s.

Solid State Disk

To modify or delete small files, SSD must delete a whole block at a time, and move content in existing blocks to a new block. SSD is fast when brand new, but will get fragmented over time and you have to garbage collect, compacting blocks. SeaweedFS is friendly to SSD since it is append-only. Deletion and compaction are done on volume level in the background, not slowing reading and not causing fragmentation.

Back to TOC

Benchmark

My Own Unscientific Single Machine Results on Mac Book with Solid State Disk, CPU: 1 Intel Core i7 2.6GHz.

Write 1 million 1KB file:

Concurrency Level:      16
Time taken for tests:   66.753 seconds
Completed requests:      1048576
Failed requests:        0
Total transferred:      1106789009 bytes
Requests per second:    15708.23 [#/sec]
Transfer rate:          16191.69 [Kbytes/sec]

Connection Times (ms)
              min      avg        max      std
Total:        0.3      1.0       84.3      0.9

Percentage of the requests served within a certain time (ms)
   50%      0.8 ms
   66%      1.0 ms
   75%      1.1 ms
   80%      1.2 ms
   90%      1.4 ms
   95%      1.7 ms
   98%      2.1 ms
   99%      2.6 ms
  100%     84.3 ms

Randomly read 1 million files:

Concurrency Level:      16
Time taken for tests:   22.301 seconds
Completed requests:      1048576
Failed requests:        0
Total transferred:      1106812873 bytes
Requests per second:    47019.38 [#/sec]
Transfer rate:          48467.57 [Kbytes/sec]

Connection Times (ms)
              min      avg        max      std
Total:        0.0      0.3       54.1      0.2

Percentage of the requests served within a certain time (ms)
   50%      0.3 ms
   90%      0.4 ms
   98%      0.6 ms
   99%      0.7 ms
  100%     54.1 ms

Run WARP and launch a mixed benchmark.

make benchmark
warp: Benchmark data written to "warp-mixed-2023-10-16[102354]-l70a.csv.zst"                                                                                                                                                                                               
Mixed operations.
Operation: DELETE, 10%, Concurrency: 20, Ran 4m59s.
 * Throughput: 6.19 obj/s

Operation: GET, 45%, Concurrency: 20, Ran 5m0s.
 * Throughput: 279.85 MiB/s, 27.99 obj/s

Operation: PUT, 15%, Concurrency: 20, Ran 5m0s.
 * Throughput: 89.86 MiB/s, 8.99 obj/s

Operation: STAT, 30%, Concurrency: 20, Ran 5m0s.
 * Throughput: 18.63 obj/s

Cluster Total: 369.74 MiB/s, 61.79 obj/s, 0 errors over 5m0s.

To see segmented request statistics, use the --analyze.v parameter.

warp analyze --analyze.v warp-mixed-2023-10-16[102354]-l70a.csv.zst
18642 operations loaded... Done!
Mixed operations.
----------------------------------------
Operation: DELETE - total: 1854, 10.0%, Concurrency: 20, Ran 5m0s, starting 2023-10-16 10:23:57.115 +0500 +05
 * Throughput: 6.19 obj/s

Requests considered: 1855:
 * Avg: 104ms, 50%: 30ms, 90%: 207ms, 99%: 1.355s, Fastest: 1ms, Slowest: 4.613s, StdDev: 320ms

----------------------------------------
Operation: GET - total: 8388, 45.3%, Size: 10485760 bytes. Concurrency: 20, Ran 5m0s, starting 2023-10-16 10:23:57.12 +0500 +05
 * Throughput: 279.77 MiB/s, 27.98 obj/s

Requests considered: 8389:
 * Avg: 221ms, 50%: 106ms, 90%: 492ms, 99%: 1.739s, Fastest: 8ms, Slowest: 8.633s, StdDev: 383ms
 * TTFB: Avg: 81ms, Best: 2ms, 25th: 24ms, Median: 39ms, 75th: 65ms, 90th: 171ms, 99th: 669ms, Worst: 4.783s StdDev: 163ms
 * First Access: Avg: 240ms, 50%: 105ms, 90%: 511ms, 99%: 2.08s, Fastest: 12ms, Slowest: 8.633s, StdDev: 480ms
 * First Access TTFB: Avg: 88ms, Best: 2ms, 25th: 24ms, Median: 38ms, 75th: 64ms, 90th: 179ms, 99th: 919ms, Worst: 4.783s StdDev: 199ms
 * Last Access: Avg: 219ms, 50%: 106ms, 90%: 463ms, 99%: 1.782s, Fastest: 9ms, Slowest: 8.633s, StdDev: 416ms
 * Last Access TTFB: Avg: 81ms, Best: 2ms, 25th: 24ms, Median: 39ms, 75th: 65ms, 90th: 161ms, 99th: 657ms, Worst: 4.783s StdDev: 176ms

----------------------------------------
Operation: PUT - total: 2688, 14.5%, Size: 10485760 bytes. Concurrency: 20, Ran 5m0s, starting 2023-10-16 10:23:57.115 +0500 +05
 * Throughput: 89.83 MiB/s, 8.98 obj/s

Requests considered: 2689:
 * Avg: 1.165s, 50%: 878ms, 90%: 2.015s, 99%: 5.74s, Fastest: 99ms, Slowest: 8.264s, StdDev: 968ms

----------------------------------------
Operation: STAT - total: 5586, 30.2%, Concurrency: 20, Ran 5m0s, starting 2023-10-16 10:23:57.113 +0500 +05
 * Throughput: 18.63 obj/s

Requests considered: 5587:
 * Avg: 15ms, 50%: 11ms, 90%: 34ms, 99%: 80ms, Fastest: 0s, Slowest: 245ms, StdDev: 17ms
 * First Access: Avg: 14ms, 50%: 10ms, 90%: 33ms, 99%: 69ms, Fastest: 0s, Slowest: 203ms, StdDev: 16ms
 * Last Access: Avg: 15ms, 50%: 11ms, 90%: 34ms, 99%: 74ms, Fastest: 0s, Slowest: 203ms, StdDev: 17ms

Cluster Total: 369.64 MiB/s, 61.77 obj/s, 0 errors over 5m0s.
Total Errors:0.

Back to TOC

License

Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.

The text of this page is available for modification and reuse under the terms of the Creative Commons Attribution-Sharealike 3.0 Unported License and the GNU Free Documentation License (unversioned, with no invariant sections, front-cover texts, or back-cover texts).

Back to TOC

Stargazers over time

Stargazers over time

seaweedfs's People

Contributors

banjiaojuhao avatar bingoohuang avatar blacktear23 avatar bukton123 avatar chrisluroblox avatar chrislusf avatar chrisluuber avatar dependabot[bot] avatar dragonstuff avatar elee1766 avatar ernado avatar famosss avatar gfxlabs avatar guo-sj avatar guol-fnst avatar hilimd avatar iliul avatar jessebot avatar kmlebedev avatar lazydba247-anyvision avatar papayofen avatar qieqieplus avatar ryanrussell avatar shichanglin5 avatar song-zhang avatar tnextday avatar woellchen avatar wusongankang avatar yanyiwu avatar yourchanges avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

seaweedfs's Issues

How can I append file

I have a file has uploaded , I want to append this file‘s content, how can do it ?

Reading Multi part error

Getting tons of these errors, unable to upload. Any ideas? Restarted the volume and things are working now.

I0115 15:02:57 00001 needle.go:65] Reading Multi part [ERROR] EOF

(Edit: restarting didn't fix things)

image id format change?

It seems like image ids have dramatically increased in size (I'm assuming since I've updated versions)... is this expected?

old style: 4,92569a439e91
new style: 4,fee4fdc8838a1ec4f305f7fe

Seems like "-log_dir" not parsed

OS: Ubuntu 14.04.1 LTS
Weed version: 0.68 (downloaded from pre-compiled deb package)

Run Command:

/usr/bin/weed -log_dir="/var/log/weed" -alsologtostderr=false -v=4 master -mdir="/tmp" -port=9333 -volumeSizeLimitMB=30000 -ip="10.130.197.44"

Viewing Log Command:

root@static:/var/log# tail /var/log/weed
tail: cannot open ‘/var/log/weed’ for reading: No such file or directory

Fix travis-ci badge

You will need to sign up to
https://travis-ci.org/
With your github account

Go to your profile
Sync with you repos
Switch "on" chrislusf/weed-fs

If you will have questions or difficulties - feel free to ask.

P.s. currently build is failing due to compile-time error

Wrong HTTP verb semantics around file creation

According to the README, the /assign request is a "GET", but performing this "GET" creates a space for a file. HTTP GET actions are supposed to only be used for actions that do not modify state (ie. they can be cached, and a spider re-requesting them thousands of times a second won't have any effect other than a transient increase in server load).

The /assign request should be a POST, as it creates something. This avoids intermediate components breaking things by making assumptions about the behavior of GET requests (ie. a caching layer not forwarding the GET request to the server).

(Reading the further API documentation, this also applies to /vol/vacuum, /vol/grow, and /admin/assign_volume. GET is an appropriate verb for /dir/lookup, /cluster/status, /dir/status, and /status.)

Similarly, the requests to upload file content are "POST" requests, but they should be "PUT" requests, as they are replayable to restore state.

ReadTheDocs documentation

Move docs to RTD and extend them with docker examples

(This issue is for me, if you want - add me to collaborators and assign it to me)

Thank you

Thanks for adding this to Github, even if it's just a mirror. I suggest moving over development here as it will help you gain help and possibly start a community.

I plan to give a helping hand at some point in the future, if time allows, and play with it on our servers. There's still a lot of work to be done to catch up with the big guys (gluster, ceph, etc). I really dig the motivation of the project, though (simple, no bullshit, key-value file storage). I hope you're successful in finding a wider user base. 👍

what about this logo?

image

My girlfriend made it.
Maybe it can be used as this project's logo if you like it?

content type from volume post upload always text/plain?

I'm assuming the return content-type should be application/json?

$ curl -v -F "file=@/tmp/test.json;type=application/json" localhost:8080/7,07a27ee211
> POST /7,07a27ee211 HTTP/1.1
> User-Agent: curl/7.30.0
> Host: localhost:8080
> Accept: */*
> Content-Length: 223
> Expect: 100-continue
> Content-Type: multipart/form-data; boundary=----------------------------2f1cf300fd2b
> 
< HTTP/1.1 100 Continue
< HTTP/1.1 201 Created
< Date: Mon, 29 Sep 2014 20:08:45 GMT
< Content-Length: 30
< Content-Type: text/plain; charset=utf-8
< 
* Connection #0 to host localhost left intact
{"name":"test.json","size":74}

File Listing or Finding

Is there a way to find or list the files? Is there a simple plan so I can code it myself? I do like the simplicity of this.

Missing build instructions for Mac OSX 10.9.4

I can't find the precompiled binary release for my OS platform (Mac OSX 10.9.4) on your linked page.

When I git clone this repo, I find a bunch of *.go files in some subfolders. Is there a Makefile to compile all of these into a weed binary myself?

Volume server keep wanting only to talk to localhost:9333

I am trying to get volume servers to talk to a master but no mater what I say on the command arguments for mserver it starts sticking trying to connect to localhost:

Starting volume server(on IP 10.150.14.80):
/usr/local/bin/weed/weed -log_dir=/var/log/weed -v=4 -alsologtostderr=true volume -dir=/data -mserver="10.150.14.89:9333"

Starting master server(ON IP 10.150.14.89)
/usr/local/bin/weed/weed -v=4 master -mdir=/data

The log on the volume server keeps saying:
I1106 22:00:14 04343 file_util.go:20] Folder /data Permission: -rwxr-xr-x
I1106 22:00:14 04343 store.go:220] Store started on dir: /data with 0 volumes max 7
I1106 22:00:14 04343 volume.go:91] Start Seaweed volume server 0.65 at 0.0.0.0:8080
I1106 22:00:14 04343 list_masters.go:18] list masters result :{"IsLeader":true,"Leader":"localhost:9333"}
I1106 22:00:14 04343 store.go:57] current master node is :localhost:9333
I1106 22:00:14 04343 volume_server.go:65] Volume Server Failed to talk with master: Post http://localhost:9333/dir/join: dial tcp 127.0.0.1:9333: connection refused

Any ideas ?

export的命令行参数与实现是否不一致?

//export.go

exportVolumeId   = cmdExport.Flag.Int("volumeId", -1, "a volume id. The volume should already exist in the dir. The volume index file should not exist.")

the flag parameter says : "The volume index file should not exist"

indexFile, err := os.OpenFile(path.Join(*exportVolumePath, fileName+".idx"), os.O_RDONLY, 0644)

but at here, open the index file with flag: O_RDONLY

is it a misspelling ?

Public port & url

Use case:
I have a weed servers:

s1.hostname.com:8080
s2.hostname.com:8080

I can place them behind reverse-proxy (nginx), and they will be accessible as

hostname.com/media/s1/:80
hostname.com/media/s2/:80

or even

hostname.com/media:80

or even

hostname.com/media:443

Which is

http://hostname.com/media
https://hostname.com/media

In case of 100 replication enabled on master server.

Now i can achieve this only with manual rewrite rules for s1 and s2.

It could be like

weed volume -port 8080 -publicPort 80 -publicPort 443 -publicURL "hostname.com/media"

Or something like this

Docker build not working

Seems like it can't opkg-install curl

$ docker build -t weedfs .
Sending build context to Docker daemon 50.78 MB
Sending build context to Docker daemon 
Step 0 : FROM progrium/busybox
Pulling repository progrium/busybox
6f114d3139e3: Download complete 
511136ea3c5a: Download complete 
e2fb46397934: Download complete 
015fb409be0d: Download complete 
21082221cb6e: Download complete 
bbd692fe2ca1: Download complete 
8db8f013bfca: Download complete 
7fff0c6f0b8d: Download complete 
7acf13620725: Download complete 
Status: Downloaded newer image for progrium/busybox:latest
 ---> 6f114d3139e3
Step 1 : WORKDIR /opt/weed
 ---> Running in 028351e07e71
 ---> d1d57d22ef4d
Removing intermediate container 028351e07e71
Step 2 : RUN opkg-install curl
 ---> Running in 2fd9079c8d42
wget: bad address 'downloads.openwrt.org'
wget: bad address 'downloads.openwrt.org'
Downloading http://downloads.openwrt.org/snapshots/trunk/x86_64/packages/base/Packages.gz.
Downloading http://downloads.openwrt.org/snapshots/trunk/x86_64/packages/packages/Packages.gz.
Collected errors:
 * opkg_download: Failed to download http://downloads.openwrt.org/snapshots/trunk/x86_64/packages/base/Packages.gz, wget returned 1.
 * opkg_download: Failed to download http://downloads.openwrt.org/snapshots/trunk/x86_64/packages/packages/Packages.gz, wget returned 1.
Unknown package 'curl'.
Collected errors:
 * opkg_install_cmd: Cannot install package curl.
 ---> 958aa12283bc
Removing intermediate container 2fd9079c8d42
Step 3 : RUN echo insecure >> ~/.curlrc
 ---> Running in 576c9f7a23d3
 ---> c39d8579e570
Removing intermediate container 576c9f7a23d3
Step 4 : RUN curl -Lks https://bintray.com$(curl -Lk http://bintray.com/chrislusf/Weed-FS/seaweed/_latestVersion | grep linux_amd64.tar.gz | sed -n "/href/ s/.*href=['\"]\([^'\"]*\)['\"].*/\1/gp") | gunzip | tar -xf - -C /opt/weed/ &&   mv weed_* bin &&   chmod +x ./bin/weed
 ---> Running in 5d84f61c0e11
/bin/sh: curl: not found
/bin/sh: curl: not found
gunzip: invalid magic
tar: short read
INFO[0095] The command [/bin/sh -c curl -Lks https://bintray.com$(curl -Lk http://bintray.com/chrislusf/Weed-FS/seaweed/_latestVersion | grep linux_amd64.tar.gz | sed -n "/href/ s/.*href=['\"]\([^'\"]*\)['\"].*/\1/gp") | gunzip | tar -xf - -C /opt/weed/ &&   mv weed_* bin &&   chmod +x ./bin/weed] returned a non-zero code: 1 
$ docker --version
Docker version 1.4.1, build 5bc2ff8

go fmt & golint

I think it will be awesome to solve some golint warnings/proposals and make go fmt on all code.

How to backup data?

How can I backup weed data? Suppose I can't stop writes. Can you explain it and add it to docs?

get content type error

hi,

I use weed-fs 1 years ago. It is an old version [0.45] of weed.
I post an image to weedfs, then I get it from browser , It will display correctly.

curl -F file=@/home/chris/myphoto.jpg http://localhost:9333/submit

when I use 0.67 of weed. I do the same post, But the same images cannot display in the browser correctly. It cannot display then do download in the chrome.

what is wrong about the weed version up? How can I config to this get header output ?

please help me.

The replication in weed is Strong consistency?

I saw the codes in ReplicatedWrite (store_replicate.go) ,
When the file uploaded, ReplicatedWrite will do replication (to another volume server) first, and when the replication is done, return the result , so I think The replication in weed is Strong consistency , is it right?

single point of failure

3 master server: mA mB mC
2 volume server: vA vB

mA is the leader of (mA,mB,mC) cluster.
and vA and vB are connected with mA; (replication is "001")

If mA is down, mB become the leader of (mB,mC) as my expectation.
But why don't vA and vB is still looking up for mA when writing replication?
When vA and vB will realize to looking up for mB for writing replication?

In this case, when mA is down and then I submit file to mB;
it always returns e.g. {"error":"Failed to write to replicas for volume 3"}

How to solve this problem about SPOF?

Add document into github

Please Add/Copy document from googlecode into github page or wiki. We no need to peeping googlecode.

Weed "minimal" docker image

I think the weed docker image which build from the given Dockerfile is too large.

update: I made a pull request

-filer.redirectOnRead should imply -filer

If I start a 'weed server -filer.redirectOnRead' without specifying -filer, it gives a puzzling error message because the filer does not start. If the user sets -filer.redirectOnRead, make it start the filer too.

weed listen ip on 0.0.0.0 ,regardless of -ip= setting

this is a bug ? find in 0.67-0.68 version

[root@backgroup ~]# ps aux | grep weed
root 15526 0.0 0.0 103252 840 pts/0 R+ 17:23 0:00 grep weed
weedfs 32554 0.0 0.0 777940 24076 ? Sl 2014 21:58 /usr/local/weed/weed master -mdir=/opt/weedfs -ip=10.168.241.178
[root@backgroup ~]# netstat -tnlp
Active Internet connections (only servers)
Proto Recv-Q Send-Q Local Address Foreign Address State PID/Program name
tcp 0 0 0.0.0.0:6379 0.0.0.0:* LISTEN 11159/redis-server
tcp 0 0 0.0.0.0:11211 0.0.0.0:* LISTEN 4755/memcached
tcp 0 0 0.0.0.0:4369 0.0.0.0:* LISTEN 4621/epmd
tcp 0 0 0.0.0.0:9333 0.0.0.0:* LISTEN 32554/weed

How to migrate volume from server `A` to `B`?

Suppose we've got 3 volume servers (A, B, C) with 001 replication type. We wan't to remove volume server C. We have to migrate all volumes located on volume server C to A and B.

Is it possible to migrate volume with id X from server A to server B?

expose http GET only

Is there currently a way to expose the GET resource separately from the other GETs & POSTs?

ie:

I would like to (publicly) provide access to

GET http://server1.example.com:8080/3,238ceaa9a7

but not to

POST http://server1.example.com:8080/3,238ceaa9a7

nor

GET http://server1.example.com:8080/status (etc...)

Right now my alternatives seem to be:

  • iptable string matching (which doesn't seem very fun)
  • proxy server (extra hop + defeats the purpose of fully concurrent reads)

I think a reasonable solution would be to provide some alternative ip:port binding for just the resource GETs.

Volume/ Master not linking

I have the following server configuration in DigitalOcean:

  • 2x Ubuntu 14.04.1 LTS
  • Local network IP enabled
  • Weedfs installed from "weed_0.68_amd64.deb" package

Server 1 (Master) IP: 10.130.197.44
Server 2 (Volume) IP: 10.130.161.166

Server 1 run command:

root@static:~# weed master -mdir="/tmp" -port=9333 -volumeSizeLimitMB=30000 -ip="10.130.197.44" -whiteList="127.0.0.1,10.130.161.166"
I0121 21:30:10 05111 file_util.go:20] Folder /tmp Permission: -rwxrwxrwx
I0121 21:30:10 05111 topology.go:86] Using default configurations.
I0121 21:30:10 05111 master_server.go:58] Volume Size Limit is 30000 MB
I0121 21:30:10 05111 master.go:70] Start Seaweed Master 0.68 at 0.0.0.0:9333
I0121 21:30:10 05111 raft_server.go:103] Old conf,log,snapshot should have been removed.

Server 2 run command:

root@vol1:~# weed volume -dir="/var/datastore" -ip="10.130.161.166" -max=7 -port=8080 -mserver="10.130.197.44:9333"
I0121 21:31:23 01110 file_util.go:20] Folder /var/datastore Permission: -rwxr-xr-x
I0121 21:31:23 01110 store.go:232] Store started on dir: /var/datastore with 0 volumes max 1
I0121 21:31:23 01110 volume.go:94] Start Seaweed volume server 0.68 at 0.0.0.0:8080

Been waiting for 10 mins and both console remains the same, seems to be stuck.

redirectOnRead goes to localhost

I've got it running as a server:

./weed server -filer.redirectOnRead -filer

Here's the curl. Note the hostname in the Location header:

bash-3.2$ curl -vI http://weed1:8888/path/to/sources/README.md > /dev/null

HEAD /path/to/sources/README.md HTTP/1.1
User-Agent: curl/7.30.0
Host: weed1:8888
Accept: /

< HTTP/1.1 302 Found
< Location: http://localhost:8080/5,01afbd8c36
< Date: Sat, 20 Dec 2014 17:15:34 GMT
< Content-Type: text/plain; charset=utf-8
<
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0

  • Connection #0 to host weed1 left intact
    bash-3.2$

I think weed-fs needs a web console

For ops can see the weed-fs running information or some helpful command.

For example, yesterday I met a problem about two master server both thinking itself is the leader,
so that clients can not upload any file (by /submit ) .
And if it has a web console, anyone can easily look up the cluster information any time to find out the problem.

In the end, we restarted the two master server to solve this problem.

Restoring from idx and dat only?

I'm migrating volumes from one machine to another.

It seems like volumes start up fine with only idx and dat files. (I see the volumes being loaded in the logs.)

When the volumes are connected to a fresh master and I do /dir/assign, the master grows new volumes instead of allocating into the existing volumes. Is this expected behavior?

benchmark not working properly

on my mac, call ./weed benchmark -server=localhost:9333

then got the following, it seems stuck at "Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s" and does not exit

Completed 964638 of 1048576 requests, 92.0% 5197.2/s 5.2MB/s
Completed 970089 of 1048576 requests, 92.5% 5452.4/s 5.5MB/s
Completed 976989 of 1048576 requests, 93.2% 6889.6/s 6.9MB/s
Completed 984025 of 1048576 requests, 93.8% 7034.8/s 7.1MB/s
Completed 991143 of 1048576 requests, 94.5% 7118.4/s 7.2MB/s
Completed 998137 of 1048576 requests, 95.2% 6979.6/s 7.0MB/s
Completed 1005152 of 1048576 requests, 95.9% 7023.3/s 7.1MB/s
Completed 1012222 of 1048576 requests, 96.5% 7069.1/s 7.1MB/s
Completed 1019163 of 1048576 requests, 97.2% 6938.4/s 7.0MB/s
Completed 1026164 of 1048576 requests, 97.9% 6989.8/s 7.0MB/s
Completed 1033175 of 1048576 requests, 98.5% 7026.3/s 7.1MB/s
Completed 1039418 of 1048576 requests, 99.1% 6240.1/s 6.3MB/s
Completed 1046381 of 1048576 requests, 99.8% 6942.4/s 7.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 1803.5/s 1.8MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s
Completed 1048181 of 1048576 requests, 100.0% 0.0/s 0.0MB/s

Backups / Migrations

Hi,

I'd like to migrate data from a running weed-fs to another one (from a staging server to a production one for instance), or even just make a backup of the files (that I could re-import if needed, not just a big dump)... is there some export/import/migration possibility in place ?

Thanks!

Can I hide port info from volume publicUrl?

We would like to proxy requests over nginx. So even if volume server accepts requests on 8080 our clients will use 80 port. I was able to change volume publicUrl by -publicIp= option but it still returns the same (8080) port. Is it possible to hide port for publicUrl?

Stability of master servers

I'm using weed-fs in production now. Sometimes master server randomly respond with blank line or "no more volumes" error. (even when correct topology in /dir/status)
I used 3 master-servers and it caused to even decrease overall stability of the system - sometimes master servers lost sync with each other / unable to promote the leader, and the only way fixing it was removing all master server files, shutting down all master servers and data nodes, then starting master servers and nodes.

I'm not quite sure how to investigate this problem, because reproducibility is low - I have no idea how to force master server(s) into unavailable state.

Maybe some hardcore unit/integration tests?
This issue not allowing me to use weed-fs in production for my new projects and I'm currently migrating to cloud storage, but leaving possibility for migrating back.

Add a tool to change a volume's replication setting.

Hello. We are going to try weed-fs in our pretty big project to store 3m+ images. There are few questions about replication which are not clear from docs:

  1. Weed replicates volumes. Is it correct?
  2. Is it possible to change replication type for volume x? Suppose we added one more server
  3. Is it possible to control volume location? We've got 3 servers and 001 replication type. Volume 1 is located on server A and B. How can I move it from B to C?

-filer.dir perms

I'm trying to start the server without filer by adding "-filer=false" but it still asks for the permissions. is that logical issue?

 Check Mapping Meta Folder (-filer.dir="") Writable: stat : no such file or directory
goroutine 16 [running]:

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.