Discover reviews on "where to host ollama" based on Reddit discussions and experiences.
Last updated: December 8, 2024 at 02:51 PM
Summary of Reddit Comments on "Where to Host Ollama"
Hosting Options
- GPU Instances can be very expensive, so some users opt for more cost-effective options like Runpod.
- LocalLlama folks often use Pascal series GPUs like P100 and P40 which can be obtained for around $150 each.
- For a pay per token API supporting OpenAI API, users recommend DeepInfra or OpenRouter as cost-effective options.
- Digital Ocean is suggested for hosting using Open Web-UI as the GUI.
- Users mention options like hosting at home and using ngrok or setting up a localllama setup to send it for Colocation.
- Deterrence from using an LLM hosted where you pay to use the API like gpt-4 is highlighted.
- A user shares the smallest and cheapest RAG setup on GCP for about $50 a month.
Cheaper Options and Small Projects
- Pinecone.io is recommended for prototyping or small projects due to its serverless free option.
- OpenAI and services like Heroku are praised for being cost-effective with a monthly cost of around $12.
- Fly.io is proposed as an option that users might find appealing.
Hosting with Other AI Tools
- A super easy way to host an Ollama instance on DigitalOcean is suggested with a provided link.
- Users advise against using more expensive GPU instances and suggest alternatives like Serverless options.
User Feedback on AI Projects
- Users commend the effort put into AI projects and the positive impact they can have on research and understanding.
- Some users appreciate the creative and informative nature of AI projects and encourage continued work in this field.
Comments on Building and Managing Ollama
- Some users share their experiences and setups for building and running Ollama instances.
- There is a mix of opinions on hardware choices, VRAM requirements, and suggestions for optimizing performance for running LLMs.
AI Chat Clients
- Various AI chat clients like BoltAI, MindMac, OpenCat, Machato are reviewed and compared based on features and performance.
- Users discuss sharing feedback with developers, suggesting improvements like a floating window feature for better multitasking capabilities.
- Some users express difficulty in accessing or using specific AI chat apps based on availability or compatibility.
Other Discussions
- There are discussions on the practicality and cost-effectiveness of self-hosting, concerns about data security, and comparisons of various hosting solutions and AI tools available.
- Users share their experiences with self-hosting services and the considerations they make in terms of cost, maintenance, and performance.
- Overall, the comments provide a range of insights and feedback on different aspects of hosting and using AI services like Ollama.