-
Notifications
You must be signed in to change notification settings - Fork 188
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How do you use LizardFS? #858
Comments
I am using Lizardfs in two SOHO setups I don't use lizardfs as my main home storage, as I was not able to go around boosting the write speeds with SSD with an automatic move to colder HDD storage after a while (as offered by other solutions) |
We use lizardfs to store millions of small files, through our web service we must have them available to our customers.
initially that worked well the first months, but then we started having problems with the metadata servers and we had to remove URAFT and later remove lizardfs from our productive environment due to the high consumption of RAM in the metadata servers (81GB RAM of 94GB) and due to problems when starting the services (I have a pending thread here #824 ) We managed to store 207,000,000 small files with a total storage of 19TB and 15TB used. We gave another opportunity to lizardfs only using servers in Datacenter A and without URAFT. It currently has 77 million small files stored, our cluster consists of
At the moment we have been with this cluster for approximately 4 months and we have not had any problem. |
We have several clusters all running in our datacenter: The first is the oldest, and we're migrating everything off of it because of its age: LFS 3.9.4 LFS stores KVM VMs, which all three run with live migration capabilities. It also stores all of the user data. The VMs are small (usually about 8GB) and just run the OS and any installed applications, but then mount various LFS mounts for user data storage. The VMs provide services for web, email, VoIP, and a few other hosted services. The second cluster is running 3.13rc1, similar to the above setup, except with uRaft, and Proxmox vs. KVM. Both of those clusters are all running SSDs only. The third cluster is for larger slow storage running 3.13rc2 and is using HDDs. All servers are connected via 10GbE. |
We have a cluster which has been running for 3 years now using LFS 3.12.0 The config is as follows: Spread over the chunk servers we have 37 spinning drives all 3-4TB enterprise grade. In total we have 13T used of 108TB in 32,889,057 files and 32,801,350 chunks
We have stopped adding data to the system as the performance is terrible and we are currently waiting for LizardFS updates or finding another solution. One of the projects on the system will be leaving soon so we can remove millions of files and hopefully the system will be usable enough that we can test performance enhancements without impacting clients. |
@HobbledGrubs This is odd also for me. Version you are currently using should be stabile. Could you please share with me more details about your problem via mail: pawel.pinkosz@lizardfs.com? |
Let's keep as much information public as possible please. I'm also interested to know more about the problem and solution... |
Ok, here is a resume of what happened.
Here are some logs:
A list of all segfaults
chunk02 was the server rebooted at Feb 5 17:24:31 |
I agree that it would be good to keep the discussion about the problem public, but we should probably create a new issue for it and then link to it from here so that we don't over-clutter this topic. |
We use LizardFS for content storage for television delivery. 3.12 7 Servers: 2x5TB, 2X15TB, 2x22TB 1x44TB, for 127TB, with ~8TB free. Expected to grow. A 3 server Proxmox cluster holds master, and a shadow, cgi-serv, this aids in the event of needing to do maintenance, snapshots. 10G backbone is planned, but missing cards. We use Debian, XFS, mostly Areca sata controllers. Our DAM now has support for managing Goals dynamically, changes to command line clients commands, would introduce changes for us. Nothing horrible, but knowing about the changes in advance would be helpful. RMlint is used for de-duplication using extended file attributes for hash storage, we also have support in out DAM to use/create this same (md5) hash. Any changes that might break extended file attributes would be unfortunate. Clients are Debian and OSX, that windows client, or build instructions, would be "handy" I had some issues with a 2 chunk goal ending with odd missing chunks when losing a drive. Try to keep 3 chunks and or parity now if the content matters. Other than that it's been pretty good, MUCH easier to manage than CEPH. |
I'm using it on a single node (my home server) with 5 chunkservers (1 for each hdd) so I can set different profiles for different files. General media gets xor3, my photos get copies 3, personal files get copies 2, etc. It works pretty well until I try to read to many files too quickly, then I run out of ram (12gb total) and lizard crashes until I reboot my machine. Strange things happen when lizard crashes while mounted... |
We have three setup, second is on a IDC, 1 master 1 slave, 10 chunkservers, total 28T and mostly big backup files, and a few exports to web server for user upload storage. All our lizard server are running on bare metal boxes with direct hard drivers, but is not the only service on those box. Those box serve as Proxmox nodes. We also has about 100 millions small file need to storage, we choose SeaweedFS as it's design for small files, with only master and volume server (we save fid with our application data), it's stable enough for our needs. |
Hi guys, if anyone would like to share your story with us – please email me antonio@lizardfs.com. We are looking for any use case of LizardFS. Your input into the project would be much appreciated. E-mail me and I'll tell you which way it will be appreciated :) |
Sorry to say, I don't use LizardFS any more. Unfortunately the 3.13 RC1 stuff completely turned me off, especially when combined with the poor performance on my ODROID HC2 units... I went back to MooseFS... I'd happily explore LizardFS again if there was any hope of it running as fast as MooseFS on my hardware. |
@Zorlin I am sure we will make you change your mind soon :) Thank you for your feedback though. |
We use lizardFS as massive cache distribution layer.
Our students primarily use VMware Workstation and work on templates that are provided on a central nfs storage. Doing this, all our templates are distributed to all the clients, which mount the share. Having PREFER LOCAL CHUNKSERVER enabled, each client will sip the templates from his own SSD if the chunks in question are available. If unsuccessful, they will try to get the chunk from a node on the same switch, if unsuccessful, it will try to get them from ANY other node on the network running lizardFS. This turned out to work extremely well, performance-, availability- and administrationwise. |
Placeholder for future explanation of still in service 2.6.0-wip +(Corosync/Pacemaker). Three LFS clusters do iSCSI/NFS block stores servicing multiple Xen pools. Eight LFS clusters do mixed native and block stores. Mixes of Five ~100TB JBOD and Five more smaller JBOD/Raid bare metal CS. Various SSD/NVMe bare and virtual CS. Masters/Shadows a mix of metal and mostly virtual. Loggers at CSs, Day/weekly snapshots, no dedup (yet) but just read someone (#858 (comment)) is, was planning on the same tricks. Running metadata sets sum up to about ~300GB, largest is kept under 128G RAM. Due to network glitches mostly, often leave CoroPace inactive but still use it to transition and move FloatIP between masters. Standard goals, never used EC, no long haul replication yet. One day, will get to 10-100gbps network. At home: ODroid/RasPi/VirtualBox and (10+16TB USB)³, one CS is even on a laptop via CIFS+VirtualBox. Mainly block store of decades worth of junk. Backs some Sia storage hosts and chewing on Ring footage as of late. |
Last configuration. Two buildings with 5 servers each 5x 1080 TB per site. Running HA and topology. The user downloads from the nearest location. |
[Note: all LizardFS clusters are running Ubuntu-MATE 22.04 LTS, with Ubuntu-MATE 20.04 LTS lizardfs*.deb's] |
I just wanted to open a general topic asking how the people that use LizardFS are using it today. It would be valuable to know things like how big your cluster is ( number of nodes, cluster storage capacity ), what hardware you are deploying to ( VMs, bare metal, SSD, etc. ), and if you are using any deployment tools such as Docker, Ansible, Chef, etc. Possibly another good thing to know would be how you plan to use LizardFS in the future, if at all.
Obviously nobody is obliged to answer, but I think that it would help us get an idea of what improvements to target with LizardFS if we knew how people use it today.
To answer my own questions, I'm not using LizardFS at the moment.
In the past I used my team's Docker image and Docker plugin to deploy LizardFS on Docker Swarm. This was a 3 node cluster where the apps ran on the same servers as LizardFS. The servers were VMs in the cloud with SSDs.
The cluster hosted database storage and the file storage for all of the apps we hosted on those servers. Databases were noticeably slower on LizardFS, but it worked. We were running LizardFS 3.12.
With LizardFS development getting back up again, me and my team very possibly will use it in the future, but we don't have a specific use-case in mind yet.
The text was updated successfully, but these errors were encountered: