Reddit User Feedback on Major LLM Chat Tools
Overview: This report analyzes Reddit discussions about four popular AI chat tools – OpenAI’s ChatGPT, Anthropic’s Claude, Google’s Gemini (Bard), and open-source LLMs (e.g. LLaMA-based models). It summarizes common pain points users report for each, the features they most frequently request, unmet needs or user segments that feel underserved, and differences in perception among developers, casual users, and business users. Specific examples and quotes from Reddit threads are included to illustrate these points.
ChatGPT (OpenAI)
Common Pain Points and Limitations
-
Limited context memory: A top complaint is ChatGPT’s inability to handle long conversations or large documents without forgetting earlier details. Users frequently hit the context length limit (a few thousand tokens) and must truncate or summarize information. One user noted “increasing the size of the context window would be far and away the biggest improvement… That’s the limit I run up against the most”. When the context is exceeded, ChatGPT forgets initial instructions or content, leading to frustrating drops in quality mid-session.
-
Message caps for GPT-4: ChatGPT Plus users lament the 25-message/3-hour cap on GPT-4 usage (a limit present in 2023). Hitting this cap forces them to wait, interrupting work. Heavy users find this throttling a major pain point.
-
Strict content filters (“nerfs”): Many Redditors feel ChatGPT has become overly restrictive, often refusing requests that previous versions handled. A highly-upvoted post complained that “pretty much anything you ask it these days returns a ‘Sorry, can’t help you’… How did this go from the most useful tool to the equivalent of Google Assistant?”. Users cite examples like ChatGPT refusing to reformat their own text (e.g. login credentials) due to hypothetical misuse. Paying subscribers argue that “some vague notion that the user may do 'bad' stuff… shouldn’t be grounds for not displaying results”, since they want the model’s output and will use it responsibly.
-
Hallucinations and errors: Despite its advanced capability, ChatGPT can produce incorrect or fabricated information with confidence. Some users have observed this getting worse over time, suspecting the model was “dumbed down.” For instance, a user in finance said ChatGPT used to calculate metrics like NPV or IRR correctly, but after updates “I am getting so many wrong answers… it still produces wrong answers [even after correction]. I really believe it has become a lot dumber since the changes.”. Such unpredictable inaccuracies erode trust for tasks requiring factual precision.
-
Incomplete code outputs: Developers often use ChatGPT for coding help, but they report that it sometimes omits parts of the solution or truncates long code. One user shared that ChatGPT now “omits code, produces unhelpful code, and just sucks at the thing I need it to do… It often omits so much code I don’t even know how to integrate its solution.” This forces users to ask follow-up prompts to coax out the rest, or to manually stitch together answers – a tedious process.
-
Performance and uptime concerns: A perception exists that ChatGPT’s performance for individual users declined as enterprise use increased. “I think they are allocating bandwidth and processing power to businesses and peeling it away from users, which is insufferable considering what a subscription costs!” one frustrated Plus subscriber opined. Outages or slowdowns during peak times have been noted anecdotally, which can disrupt workflows.
Frequently Requested Features or Improvements
-
Longer context window / memory: By far the most requested improvement is a larger context length. Users want to have much longer conversations or feed large documents without resets. Many suggest expanding ChatGPT’s context to match GPT-4’s 32K token capability (currently available via API) or beyond. As one user put it, “GPT is best with context, and when it doesn’t remember that initial context, I get frustrated… If the rumors are true about ️context PDFs, that would solve basically all my problems.” There is high demand for features to upload documents or link personal data so ChatGPT can remember and reference them throughout a session.
-
File-handling and integration: Users frequently ask for easier ways to feed files or data into ChatGPT. In discussions, people mention wanting to “copy and paste my Google Drive and have it work” or have plugins that let ChatGPT directly fetch context from personal files. Some have tried workarounds (like PDF reader plugins or linking Google Docs), but complained about errors and limits. A user described their ideal plugin as one that “works like Link Reader but for personal files… choosing which parts of my drive to use in a conversation… that would solve basically every problem I have with GPT-4 currently.”. In short, better native support for external knowledge (beyond the training data) is a popular request.
-
Reduced throttling for paid users: Since many Plus users hit the GPT-4 message cap, they call for higher limits or an option to pay more for unlimited access. The 25-message limit is seen as arbitrary and hindering intensive use. People would prefer a usage-based model or higher cap so that long problem-solving sessions aren’t cut short.
-
“Uncensored” or custom moderation modes: A segment of users would like the ability to toggle the strictness of content filters, especially when using ChatGPT for themselves (not public-facing content). They feel a “research” or “uncensored” mode – with warnings but not hard refusals – would let them explore more freely. As one user noted, paying customers see it as a tool and believe “I pay money for [it].” They want the option to get answers even on borderline queries. While OpenAI has to balance safety, these users suggest a flag or setting to relax policies in private chats.
-
Improved factual accuracy and updates: Users commonly ask for more up-to-date knowledge and fewer hallucinations. ChatGPT’s knowledge cutoff (September 2021 in earlier versions) was a limitation often raised on Reddit. OpenAI has since introduced browsing and plugins, which some users leverage, but others simply request the base model be updated more frequently with new data. Reducing obvious errors – especially in domains like math and coding – is an ongoing wish. Some developers provide feedback when ChatGPT errs in hopes of model improvement.
-
Better code outputs and tools: Developers have feature requests such as an improved code interpreter that doesn’t omit content, and integration with IDEs or version control. (OpenAI’s Code Interpreter plugin – now part of “Advanced Data Analysis” – was a step in this direction and received praise.) Still, users often request finer control in code generation: e.g. an option to output complete, unfiltered code even if it’s long, or mechanisms to easily fix code if the AI made an error. Basically, they want ChatGPT to behave more like a reliable coding assistant without needing multiple prompts to refine the answer.
-
Persistent user profiles or memory: Another improvement some mention is letting ChatGPT remember things about the user across sessions (with consent). For example, remembering one’s writing style, or that they are a software engineer, without having to restate it every new chat. This could tie into API fine-tuning or a “profile” feature. Users manually copy important context into new chats now, so a built-in memory for personal preferences would save time.
Underserved Needs or User Segments
-
Researchers and students with long documents: People who want ChatGPT to analyze lengthy research papers, books, or large datasets feel underserved. The current limits force them to chop up text or settle for summaries. This segment would benefit greatly from larger context windows or features to handle long documents (as evidenced by numerous posts about trying to get around token limits).
-
Users seeking creative storytelling or role-play beyond limits: While ChatGPT is often used for creative writing, some story-tellers feel constrained by the model forgetting early plot points in a long story or refusing adult/horror content. They turn to alternative models or hacks to continue their narratives. These creative users would be better served by a version of ChatGPT with longer memory and a bit more flexibility on fictional violence or mature themes (within reason). As one fiction writer noted, when the AI loses track of the story, “I have to remind it of the exact format or context… I get frustrated that it was great two prompts ago, but now I have to catch the AI up.”.
-
Power users and domain experts: Professionals in specialized fields (finance, engineering, medicine) sometimes find ChatGPT’s answers lacking depth or accuracy in their domain, especially if the questions involve recent developments. These users desire more reliable expert knowledge. Some have tried fine-tuning via the API or custom GPTs. Those who cannot fine-tune would appreciate domain-specific versions of ChatGPT or plugins that embed trusted databases. In its default form, ChatGPT may underserve users who need highly accurate, field-specific information (they often have to double-check its work).
-
Users needing uncensored or edge-case content: A minority of users (hackers testing security scenarios, writers of extreme fiction, etc.) find ChatGPT’s content restrictions too limiting for their needs. They are currently underserved by the official product (since it explicitly avoids certain content). These users often experiment with jailbreaking prompts or use open-source models to get the responses they want. This is a deliberate gap for OpenAI (to maintain safety), but it means such users look elsewhere.
-
Privacy-conscious individuals and enterprises: Some users (especially in corporate settings) are uncomfortable sending sensitive data to ChatGPT due to privacy concerns. OpenAI has policies to not use API data for training, but the ChatGPT web UI historically did not offer such guarantees until an opt-out feature was added. Companies that handle confidential data (legal, healthcare, etc.) often feel they cannot fully utilize ChatGPT, leaving their needs underserved unless they build self-hosted solutions. For example, a Redditor mentioned their company moving to a local LLM for privacy reasons. Until on-prem or private instances of ChatGPT are available, this segment remains cautious or uses smaller specialist vendors.
Differences in Perception by User Type
-
Developers/Technical Users: Developers tend to be both some of ChatGPT’s biggest advocates and harshest critics. They love its ability to explain code, generate boilerplate, and assist in debugging. However, they keenly feel its limitations in longer context and code accuracy. As one dev complained, ChatGPT started “producing unhelpful code” and omitting important parts, which “pisses me off… I don’t want to have to tell it ‘don’t be lazy’ – I just want the full result”. Devs often notice even subtle changes in quality after model updates and have been very vocal on Reddit about perceived “nerfs” or declines in coding capability. They also push the limits (building complex prompts, chaining tools), so they crave features like expanded context, fewer message caps, and better integration with coding tools. In summary, developers value ChatGPT for speeding up routine tasks but are quick to point out errors in logic or code – they view it as a junior assistant that still needs oversight.
-
Casual/Everyday Users: More casual users – those asking for general knowledge, advice, or fun – often marvel at ChatGPT’s capabilities, but they have their own gripes. A common casual-user frustration is when ChatGPT refuses a request that seems innocuous to them (likely tripping a policy rule). The original poster in one thread exemplified this, being “so pissed off when I write a prompt which it shouldn’t have a problem with and it refuses now”. Casual users may also run into the knowledge cutoff (finding the bot can’t handle very current events unless explicitly updated) and sometimes notice when ChatGPT gives an obviously wrong answer. Unlike developers, they might not always double-check the AI, which can lead to disappointment if they act on a mistake. On the positive side, many casual users find ChatGPT Plus’s faster responses and GPT-4’s improved output worth $20/month – unless the “refusal” issue or other limits sour the experience. They generally want a helpful, all-purpose assistant and can get frustrated when ChatGPT replies with policy statements or needs a complex prompt to get a simple answer.
-
Business/Professional Users: Business users often approach ChatGPT from a productivity and reliability standpoint. They appreciate fast drafting of emails, summaries of documents, or generation of ideas. However, they are concerned about data security, consistency, and integration into workflows. On Reddit, professionals have discussed wanting ChatGPT in tools like Outlook, Google Docs, or as an API in their internal systems. Some have noted that as OpenAI pivots to serve enterprise clients, the product’s focus seems to shift: there’s a feeling that the free or individual user experience degraded slightly (e.g. slower or “less smart”) as the company scaled up to serve larger clients. Whether or not that’s true, it highlights a perception: business users want reliability and priority service, and individual users worry they’re now second-class. Additionally, professionals need correct outputs – a flashy but wrong answer can be worse than no answer. Thus, this segment is sensitive to accuracy. For them, features like longer context (for reading contracts, analyzing codebases) and guaranteed uptime are crucial. They are likely to pay more for premium service levels, provided their compliance and privacy requirements are met. Some enterprises even explore on-premise deployments or using OpenAI’s API with strict data handling rules to satisfy their IT policies.
Claude (Anthropic)
Common Pain Points and Limitations
-
Usage limits and access restrictions: Claude received praise for offering a powerful model (Claude 2) for free, but users quickly encountered usage limits (especially on the free tier). After a certain number of prompts or a large amount of text, Claude may stop and say something like “I’m sorry, I have to conclude this conversation for now. Please come back later.” This throttling frustrates users who treat Claude as an extended coding or writing partner. Even Claude Pro (paid) users are “not guaranteed unlimited time”, as one user noted; hitting the quota still produces the “come back later” message. Additionally, for a long time Claude was officially geo-restricted (initially only available in the US/UK). International users on Reddit had to use VPNs or third-party platforms to access it, which was an inconvenience. This made many non-US users feel left out until access widened.
-
Tendency to go off-track with very large inputs: Claude’s headline feature is its 100k-token context window, allowing extremely long prompts. However, some users have noticed that when you stuff tens of thousands of tokens into Claude, its responses can become less focused. “100k is super useful but if it doesn’t follow instructions properly and goes off track, it’s not that useful,” one user observed. This suggests that with huge contexts, Claude might drift or start rambling, requiring careful prompting to keep it on task. It’s a limitation inherent to pushing context to the extreme – the model retains a lot but sometimes “forgets” which details are most relevant, leading to minor hallucinations or off-topic tangents.
-
Inconsistent formatting or obedience to instructions: In side-by-side comparisons, some users found Claude less predictable in how it follows certain directives. For example, Claude is described as “more human-like in interactions. But it less strictly follows system messages.”. This means if you give it a fixed format to follow or a very strict persona, Claude might deviate more than ChatGPT would. Developers who rely on deterministic outputs (like JSON formats or specific styles) sometimes get frustrated if Claude introduces extra commentary or doesn’t rigidly adhere to the template.
-
Content restrictions and refusals: While not as frequently criticized as ChatGPT’s, Claude’s safety filters do come up. Anthropic designed Claude with a heavy emphasis on constitutional AI (having the AI itself follow ethical guidelines). Users generally find Claude willing to discuss a broad range of topics, but there are instances where Claude refuses requests that ChatGPT might allow. For example, one Redditor noted “ChatGPT has less moral restrictions… it will explain which gas masks are better for which conditions while Claude will refuse”. This suggests Claude might be stricter about certain “sensitive” advice (perhaps treating it as potentially dangerous guidance). Another user tried a playful role-play scenario (“pretend you were abducted by aliens”) which Claude refused, whereas Gemini and ChatGPT would engage. So, Claude does have filters that can occasionally surprise users expecting it to be more permissive.
-
Lack of multimodal capabilities: Unlike ChatGPT (which, by late 2023, gained image understanding with GPT-4 Vision), Claude is currently text-only. Reddit users note that Claude cannot analyze images or directly browse the web on its own. This isn’t exactly a “pain point” (Anthropic never advertised those features), but it is a limitation relative to competitors. Users who want an AI to interpret a diagram or screenshot cannot use Claude for that, whereas ChatGPT or Gemini might handle it. Similarly, any retrieval of current information requires using Claude via a third-party tool (e.g., Poe or search engine integration), since Claude doesn’t have an official browsing mode at this time.
-
Minor stability issues: A few users have reported Claude occasionally being repetitive or getting stuck in loops for certain prompts (though this is less common than with some smaller models). Also, earlier versions of Claude sometimes ended responses prematurely or took a long time with large outputs, which can be seen as minor annoyances, though Claude 2 improved on speed.
Frequently Requested Features or Improvements
-
Higher or adjustable usage limits: Claude enthusiasts on Reddit often ask Anthropic to raise the conversation limits. They would like to use the 100k context to its fullest without hitting an artificial stop. Some suggest that even paid Claude Pro should allow significantly more tokens per day. Others floated the idea of an optional “100k extended mode” – e.g., “Claude should have a 100k context mode with double the usage limits” – where perhaps a subscription could offer expanded access for heavy users. In essence, there’s demand for a plan that competes with ChatGPT’s unlimited (or high-cap) usage for subscribers.
-
Better long-context navigation: While having 100k tokens is groundbreaking, users want Claude to better utilize that context. One improvement would be refining how Claude prioritizes information so it stays on track. Anthropic could work on the model’s prompt adherence when the prompt is huge. Reddit discussions suggest techniques like allowing the user to “pin” certain instructions so they don’t get diluted in a large context. Any tools to help segment or summarize parts of the input could also help Claude handle large inputs more coherently. In short, users love the possibility of feeding an entire book to Claude – they just want it to stay sharp throughout.
-
Plugins or web browsing: Many ChatGPT users have gotten used to plugins (for example, browsing, code execution, etc.) and they express interest in Claude having similar extensibility. A common request is for Claude to have an official web search/browsing function, so that it can fetch up-to-date information on demand. Currently, Claude’s knowledge is mostly static (training data up to early 2023, with some updates). If Claude could query the web, it would alleviate that limitation. Likewise, a plugin system where Claude could use third-party tools (like calculators or database connectors) could expand its utility for power users. This remains a feature Claude lacks, and Reddit users often mention how ChatGPT’s ecosystem of plugins gives it an edge in certain tasks.
-
Multimodal input (images or audio): Some users have also wondered if Claude will support image inputs or generate images. Google’s Gemini and OpenAI’s GPT-4 have multimodal capabilities, so to stay competitive, users expect Anthropic to explore this. A frequent request is: “Can I upload a PDF or an image for Claude to analyze?” Currently the answer is no (aside from workarounds like converting images to text elsewhere). Even just allowing image-to-text (OCR and description) would satisfy many who want a one-stop assistant. This is on the wish list, though Anthropic hasn’t announced anything similar as of early 2025.
-
Fine-tuning or customization: Advanced users and businesses sometimes ask if they can fine-tune Claude on their own data or get custom versions. OpenAI offers fine-tuning for some models (not GPT-4 yet, but for GPT-3.5). Anthropic released a fine-tuning interface for Claude 1.3 earlier, but it’s not widely advertised for Claude 2. Reddit users have inquired about being able to train Claude on company knowledge or personal writing style. An easier way to do this (besides prompt injections each time) would be very welcome, as it could turn Claude into a personalized assistant that remembers a specific knowledge base or persona.
-
Wider availability: Non-US users frequently request that Claude be officially launched in their countries. Posts from Canada, Europe, India, etc., ask when they can use Claude’s website without a VPN or when the Claude API will be open more broadly. Anthropic has been cautious, but demand is global – likely an improvement in the eyes of many would be simply “let more of us use it.” The company’s gradual expansion of access has partially addressed this.
Underserved Needs or User Segments
-
International user base: As noted, for a long time Claude’s primary user base was limited by geography. This left many would-be users underserved. For example, a developer in Germany interested in Claude’s 100k context had no official way to use it. While workarounds exist (third-party platforms, or VPN + phone verification in a supported country), these barriers meant casual international users were effectively locked out. By contrast, ChatGPT is available in most countries. So, non-US English speakers and especially non-English speakers have been underserved by Claude’s limited rollout. They may still rely on ChatGPT or local models simply due to access issues.
-
Users needing strict output formatting: As mentioned, Claude sometimes takes liberties in responses. Users who need highly structured outputs (like JSON for an application, or an answer following a precise format) might find Claude less reliable for that than ChatGPT. These users – often developers integrating the AI into a system – are a segment that could be better served if Claude allowed a “strict mode” or improved its adherence to instructions. They currently might avoid Claude for such tasks, sticking with models known to follow formats more rigidly.
-
Casual Q&A users (vs. creative users): Claude is often praised for creative tasks – it produces flowing, human-like prose and thoughtful essays. However, some users on Reddit noted that for straightforward question-answering or factual queries, Claude sometimes gives verbose answers where brevity would do. The user who compared ChatGPT and Claude said ChatGPT tends to be succinct and bullet-pointed, whereas Claude gives more narrative by default. Users who just want a quick factual answer (like “What’s the capital of X and its population?”) might feel Claude is a bit indirect. These users are better served by something like an accurate search or a terse model. Claude can do it if asked, but its style may not match the expectation of a terse Q&A, meaning this segment could slip to other tools (like Bing Chat or Google).
-
Safety-critical users: Conversely, some users who require very careful adherence to safety (e.g. educators using AI with students, or enterprise customers who want zero risk of rogue outputs) might consider Claude’s alignment a plus, but since ChatGPT is also quite aligned and has more enterprise features, those users might not specifically choose Claude. It’s a small segment, but one could argue Claude hasn’t distinctly captured it yet. They may be underserved in that they don’t have an easy way to increase Claude’s safeguards or see its “chain of thought” (which Anthropic has internally via the constitutional AI approach, but end-users don’t directly interface with that aside from noticing Claude’s generally polite tone).
-
Non-English speakers (quality of output): Claude was trained primarily on English (like most big LLMs). Some users have tested it in other languages; it can respond in many, but the quality may vary. If, say, a user wants a very nuanced answer in French or Hindi, it’s possible Claude’s abilities are not as fine-tuned there as ChatGPT’s (GPT-4 has demonstrated strong multilingual performance, often higher than other models in certain benchmarks). Users who primarily converse in languages other than English might find Claude’s fluency or accuracy slightly weaker. This segment is somewhat underserved simply because Anthropic hasn’t highlighted multilingual training as a priority publicly.
Differences in Perception by User Type
-
Developers/Tech Users: Developers on Reddit have increasingly lauded Claude, especially Claude 2 / Claude 3.5, for coding tasks. The perception shift in late 2024 was notable: many developers started preferring Claude over ChatGPT for programming assistance. They cite “amazing at coding” performance and the ability to handle larger codebases in one go. For example, one user wrote “Claude Sonnet 3.5 is better to work with code (analyze, generate) [than ChatGPT].” Developers appreciate that Claude can take a large chunk of project code or logs and produce coherent analyses or improvements, thanks to its huge context. However, they also notice its quirks – like sometimes injecting more conversational fluff or not following a spec to the letter. On balance, many devs keep both ChatGPT and Claude at hand: one for rigorous step-by-step logic (ChatGPT) and one for broad context and empathetic understanding (Claude). It’s telling that a commenter said “If I had to choose one I would choose Claude” after comparing the two daily. This indicates a very positive perception among advanced users, especially for use cases like brainstorming, code review, or architectural suggestions. The only common gripe from devs is hitting Claude’s usage limits when they try to push it hard (e.g. feeding a 50K-token prompt to analyze an entire repository). In summary, developers view Claude as an extremely powerful tool – in some cases superior to ChatGPT – held back only by availability and some unpredictability in formatting.
-
Casual/Non-technical Users: Casual users who have tried Claude often comment on how friendly and articulate it is. Claude’s style tends to be conversational, polite, and detailed. A new user comparing it to ChatGPT observed that “Claude is more empathetic, and follows a conversational tone… ChatGPT defaults to bullet points too often”. This human-like warmth makes Claude appealing to people using it for creative writing, advice, or just chatting for information. Some even personify Claude as having a “personality” that is compassionate. Casual users also like that Claude’s free version allowed access to an equivalent of GPT-4-level intelligence without a subscription (at least up to the rate limits). On the flip side, casual users do bump into Claude’s refusals on certain topics and might not understand why (since Claude will phrase it apologetically but firmly). If a casual user asked something borderline and got a refusal from Claude, they might perceive it as less capable or too constrained, not realizing it’s a policy stance. Another aspect is that Claude lacks the name recognition – many casual users might not even know to try it unless they’re tapped into AI communities. Those who do try generally comment that it feels “like talking to a human” in a good way. They tend to be very satisfied with Claude’s ability to handle open-ended or personal questions. So, casual user perception is largely positive regarding Claude’s output quality and tone, with some confusion or frustration around its availability (having to use it on a specific app or region) and occasional “can’t do that” moments.
-
Business/Professional Users: Business perceptions of Claude are a bit harder to gauge from public Reddit (since fewer enterprise users post in detail), but a few trends emerge. First, Anthropic has positioned Claude as more privacy-focused and willing to sign enterprise agreements – this appeals to companies worried about data with OpenAI. Indeed, some Reddit discussions mention Claude in the context of tools like Slack or Notion, where it’s integrated as an assistant. Professionals who have used those integrations might not even realize Claude is the engine, but when they do, they compare it favorably in terms of writing style and the ability to digest large corporate documents. For example, a team might feed a long quarterly report to Claude and get a decent summary – something ChatGPT’s smaller context would struggle with. That said, business users also notice the lack of certain ecosystem features; for instance, OpenAI offers system message control, function calling, etc., in their API, which Anthropic has more limited support for. A developer working on a business solution remarked that Claude is more steerable in conversations, whereas ChatGPT tends to be more rigid… [but] ChatGPT has web access which can be very helpful. The implication is that for research or data lookup tasks a business user might need (like competitive intelligence), ChatGPT can directly fetch info, whereas Claude would require a separate step. Overall, business users seem to view Claude as a very competent AI – in some cases better for internal analytic tasks – but perhaps not as feature-rich yet for integration. Cost is another factor: Claude’s API pricing and terms are not as public as OpenAI’s, and some startups on Reddit have mentioned uncertainty about Claude’s pricing or stability. In summary, professionals respect Claude’s capabilities (especially its reliability in following high-level instructions and summarizing large inputs), but they keep an eye on how it evolves in terms of integration, support, and global availability before fully committing to it over the more established ChatGPT.
Google Gemini (Bard)
Common Pain Points and Limitations
-
Inaccurate or “dumb” responses: A flood of Reddit feedback appeared when Google launched its Gemini-powered Bard upgrade, much of it negative. Users complained that Gemini underperformed in basic QA compared to ChatGPT. One blunt assessment titled “100% Honest Take on Google Gemini” stated: