Dark Light
Reddit Scout Logo

Reddit Scout

Discover reviews on "where to host ollama" based on Reddit discussions and experiences.

Last updated: December 8, 2024 at 02:51 PM
Go Back

Summary of Reddit Comments on "Where to Host Ollama"

Hosting Options

  • GPU Instances can be very expensive, so some users opt for more cost-effective options like Runpod.
  • LocalLlama folks often use Pascal series GPUs like P100 and P40 which can be obtained for around $150 each.
  • For a pay per token API supporting OpenAI API, users recommend DeepInfra or OpenRouter as cost-effective options.
  • Digital Ocean is suggested for hosting using Open Web-UI as the GUI.
  • Users mention options like hosting at home and using ngrok or setting up a localllama setup to send it for Colocation.
  • Deterrence from using an LLM hosted where you pay to use the API like gpt-4 is highlighted.
  • A user shares the smallest and cheapest RAG setup on GCP for about $50 a month.

Cheaper Options and Small Projects

  • Pinecone.io is recommended for prototyping or small projects due to its serverless free option.
  • OpenAI and services like Heroku are praised for being cost-effective with a monthly cost of around $12.
  • Fly.io is proposed as an option that users might find appealing.

Hosting with Other AI Tools

  • A super easy way to host an Ollama instance on DigitalOcean is suggested with a provided link.
  • Users advise against using more expensive GPU instances and suggest alternatives like Serverless options.

User Feedback on AI Projects

  • Users commend the effort put into AI projects and the positive impact they can have on research and understanding.
  • Some users appreciate the creative and informative nature of AI projects and encourage continued work in this field.

Comments on Building and Managing Ollama

  • Some users share their experiences and setups for building and running Ollama instances.
  • There is a mix of opinions on hardware choices, VRAM requirements, and suggestions for optimizing performance for running LLMs.

AI Chat Clients

  • Various AI chat clients like BoltAI, MindMac, OpenCat, Machato are reviewed and compared based on features and performance.
  • Users discuss sharing feedback with developers, suggesting improvements like a floating window feature for better multitasking capabilities.
  • Some users express difficulty in accessing or using specific AI chat apps based on availability or compatibility.

Other Discussions

  • There are discussions on the practicality and cost-effectiveness of self-hosting, concerns about data security, and comparisons of various hosting solutions and AI tools available.
  • Users share their experiences with self-hosting services and the considerations they make in terms of cost, maintenance, and performance.
  • Overall, the comments provide a range of insights and feedback on different aspects of hosting and using AI services like Ollama.
Sitemap | Privacy Policy

Disclaimer: This website may contain affiliate links. As an Amazon Associate, I earn from qualifying purchases. This helps support the maintenance and development of this free tool.