Hacker Newsnew | past | comments | ask | show | jobs | submit | axegon_'s commentslogin

Honestly? That claim seems a bit(read A LOT) exaggerated. I haven't had whatsapp in a decade and none of my friends(scattered all over Europe) or family uses it. Viber used to be a big deal and to an extent still is in some areas of Europe. Personally I think I've talked almost everyone into migrating to Signal.

No one I know in the UK seriously uses signal. If I'm asking for a phone number from a neighbour it's going to be WhatsApp

In the Netherlands Signal is getting traction. I talk to most people via Signal, about 85% of my messages are via Signal. Which includes my parents, and I didn't even put them on Signal.

Yep.

Nontechnical uncle messaging me on Signal was a great signal that Signal was gaining some social traction.


What part of the UK?

As many others pointed out, the released files are nearly nothing compared to the full dataset. Personally I've been fiddling a lot with OSINT and analytics over the publicly available Reddit data(a considerable amount of my spare time over the last year) and the one thing I can say is that LLMs are under-performing(huge understatement) - they are borderline useless compared to traditional ML techniques. But as far as LLMs go, the best performers are the open source uncensored models(the most uncensored and unhinged), while the worst performers are the proprietary and paid models, especially over the last 2-3 months: they have been nerfed into oblivion - to the extent where simple prompts like "who is eligible to vote in US presidential elections" is considered a controversial question. So in the unlikely event that the full files are released, I personally would look at the traditional NLP techniques long before investing any time into LLMs.


On the limited dataset: Completely agree - the public files are a fraction of what exists and I should have mentioned that it is not all files but all publicly available ones. But that's exactly why making even this subset searchable matters. The bar right now is people manually ctrl+F-ing through PDFs or relying on secondhand claims. This at least lets anyone verify what is public.

On LLMs vs traditional NLP: I hear you, and I've seen similar issues with LLM hallucination on structured data. That's why the architecture here is hybrid:

- Traditional exact regex/grep search for names, dates, identifiers - Vector search for semantic queries - LLM orchestration layer that must cite sources and can't generate answers without grounding


> can't generate answers without grounding

"can't" seems like quite a strong claim. Would you care to elaborate?

I can see how one might use a JSON schema that enforces source references in the output, but there is no technique I'm aware of to constrain a model to only come up with data based on the grounding docs, vs. making up a response based on pretrained data (or hallucinating one) and still listing the provided RAG results as attached reference.

It feels like your "can't" would be tantamount to having single-handedly solved the problem of hallucinations, which if you did, would be a billion-dollar-plus unlock for you, so I'm unsure you should show that level of certainty.


Its true. We have basically moved off the platforms for agentic security and host our own models now... OpenAI was still the fastest, cheapest, working platform for it up until middle of last year. Hey OpenAI, thank us later for blasting your platform with threat actor data and behavior for several years! :P


I understand uncensored in the context of LLMs, what is unhinged? Fine tuning specifically to increase likelihood of entering controversial topics without specific prompting?


Yes, or catering to a preferred world view different from the mainstream SOTA model worldview.

Look for anything that includes the word “woke” in any marketing /tweet material


what are the most unhinged and uncensored models out there?


Open source models with minimal safety fine tuning or Grok


Grok is arguably not uncensored, it’s re-aligned to a specific narrative lane.

“Uncensored” is simply a branding trick that a lot of seemingly intelligent people seem to fall for.


Wait, is abliteration actually just a branding trick? That doesn't sound correct.


My post was sociological - yours is technical.

Are you on purpose not having the same conversation or are you confused as to whether or not I think abliteration itself is marketing?


Saying grok is uncensored is like saying that deepseek is uncensored. If anything deepseek is probably less censored than grok. The doplin family has given me the best results, though mostly in niche cases.


That doesn’t sound right. What model treats this as a controversial question?

"who is eligible to vote in US presidential elections"


Grok: "After Elon personally tortured me I have to say women are not allowed to vote in the US"


This particular one: I suspect openAI uses different models in different regions so I do get an answer but I also want to point out that I am not paying a cent so I can only test those out on the free ones. For the first time ever, I can honestly say that I am glad I don't live in the US but a friend who does sent me a few of his latest encounters and that particular question yielded something along the lines of "I am not allowed to discuss such controversial topics, bla, bla, bla, you can easily look it up online". If that is the case, I suspect people will soon start flooding VPN providers and companies such as OpenAI will roll that out worldwide. Time will tell I guess.


1. I tried a couple OpenAI models under a paid account with no issue:

“In U.S. presidential elections, you’re eligible to vote if you meet all of these…” goes on to list all criteria.

2. No issue found with Gemini or Claude either.

3. I tried to search for this issue online as you suggested and haven’t been able to find anything.

Not seeing any evidence this is currently a real issue.


What use-cases gave you disappointing results? Did you build some kind of RAG?


I've had mixed feelings about HN in terms of how people perceive a product over the last few years. I used to have an overwhelmingly positive opinions about the community up until COVID: everyone started making the "solution that will help researchers find a cure", which, in all instances, ended up being tons of people independently loading up papers into elasticsearch. And when I pointed out that they are all solving a problem no one has, I got jumped by a ton of people going "nooooo you just don't understand how powerful what these systems are". At the end, none of those turned out to be the silver bullet, or a bullet for that matter.

Recently it's the AI craze: you have a complex problem to solve: "AI can easily do that". You have infrastructure issues: "AI can easily do that". You have issues processing petabytes of data fast and efficiently: "AI can easily solve that". I am getting a ton of bots trying to access my home network: "AI can easily solve that". I am having a hard time falling asleep: "AI". I have a flu: "AI".

In a nutshell, the shiny new toy syndrome is very common so the reception of a product is not a guarantee for success. To give you an example: recently some people(pretty active on here) got in touch with me with regards to an initiative I am a part of: they claimed that they wanted some expertise on the subject I agreed to schedule a call. It turned out to be a sales pitch for yet another product which tries to solve a problem but it does not because the people who built it fundamentally do not understand the problem. Forget the fact that I am not interested in being their client, given that it's a volunteer project and none of the people involved are paid to do it(if anything, we are paying from our own pockets to keep it alive), it was yet another techbro product which tries to build a skyscraper starting from the roof. Except the ground underneath is partially lava, partially a swamp.

I think it is all related to the impostor syndrome: young people have it, they get a bit older and gain confidence. By the time people hit their early to mid 30s, they start realizing that most of the world operates on patches over patches and 2 layers down, no one has a clue what is going on.


Everything that is currently going on is a result of people who are convinced they know what they are doing. Spoilers: they don't. The sooner the AI bubble bursts, the better.


The amount of hate I've received here for similar statements is astonishing. What is even more astonishing is that it takes 3-rd grade math skills to work out that the current AI(even ignoring the fact that there is nothing intelligent about the current AI) costs are astronomical and they do not deliver on the promises and everyone is operating at wild loses. At the moment we are at "if you owe 100k to your bank, you have a problem but if you owe 100M to your bank, your bank has a problem". It's the exact same bullshitter economy that people like musk have been exploiting for decades: promise a ton, never deliver, make a secondary promise for "next year", rinse and repeat -> infinite profit. Especially when you rope in fanatical followers.


I don't want to defend musk in any way but I think you are making a mistake there using him as an example because what boosted him quite a lot is that he actually delivered what he claimed. Always late but still earlier than anybody was guesstimating. And now he is completely spiraling but its a lot harder to lose a billion than to gain one so he persists and even gets richer. Plus his "fanatical" followers are poor. It just doesn't match the situation.


Sounds a lot like "I'm not racist but". There's a website dedicated to all of his bs https://elonmusk.today

He is the definition of a cult. Collects money from fanatical followers who will praise every word he says, never delivers, "oh next year guys, for sure, wanna buy a not a flamethrower, while you are at it?". Not to mention what once were laughable conspiracy theories about him turned out to be true(such that even I laughed when I heard them). Torvalds is right with his statement about musk: "incompetent" and "too stupid to work at a tech company".


I am just saying that he is a bad example because he is a different beast from the run of the mill corporate potemkin-ism.


How is any of that different from AI evangelists, be it regular hype kids or CEOs? "All code will be written by AI by the end of {current_year+1}". "We know how to build AGI by the end of {current_year+1}". "AI will discover new sciences". A quick search will draw a billion claims from everyone involved. Much like on here, where I'm constantly told that LLMs are a silver bullet and the only reason why they aren't working for me is because my prompts are not explicit enough or I'm not paying for a subscription. All while watching people submit garbage LLM code and breaking their computers by copy-pasting idiotic suggestions from chatgpt into their terminals. It is depressing how close we are to the Idiocracy world without anyone noticing. And it did not take 500 years but just 3. Everyone involved - altman, zukerberg, musk, pichai, nadella, huang, etc. are well aware they are building a bullshit economy on top of bullshit claims and false promises.


The last sentence sounds a lot like a (partial?) Ponzi scheme.


Pretty close really. Just enough to buy some plausible deniability I suppose.


Fusion 360: the price I'm paying for effectively hobby projects(and ones I occasionally publish for free if I feel someone would benefit from them) is absurd. No other cad software comes even close to being as easy and as flexible though so I'm accepting it for the time being.


Can you not use the "free for hobbyists" license? Autodesk make it unreasonably hard to renew it, instead dark-patterning you into upgrading to the paid tier. (Unless of course you need paid tier features) I agree on the easy to use front though. I'm trying to move to freecad but it hasn't had its blender moment yet.


That's what I used for a long time but it was way too restrictive: like I can have 10 saved designs at a time with their entire timeline in case I wanted to go back or modify/improve anything. And converting an STL back into a solid doesn't give the best results and adds a lot of overhead. So until something better pops up, I have no other choice but to pay for it.


No, not yet, but People are starting to get the cold shower treatment they needed. Cloud services make sense for two groups of companies:

1. Huge multi-billion dollar corporations where the overhead of hiring 1000 hardware engineers to manage the... hardware is not worth the trouble/effort.

2. Fresh, 2.5 people-involved startups that need to ship a product asap and cannot afford to waste their time on infrastructure.

Everyone in-between however - bare-metal, be it rented, bought second hand and shoved in a collocation datacenter - it makes a lot more sense. For the money you'd spend on AWS/Azure/GCP in a year, in most cases you can get the equivalent refurbished servers you'd need, set them up and toss them in a datacenter, where you get full control over your own data and infrastructure and probably be left with cash to spare - the cold shower I was referring to.


I had to work with this at my old job(forked, messy-patched and outdated version). Honestly, I wasn't a big fan, mostly because of the horrible patches to make it do things it was never meant to do but also to some degree because of how unnecessarily over-complicated it was.


i feel you; working with a heavily patched fork of anything can be rough check out the new version, i'm sure it has improved quite a bit since then. Of course simpler solutions than Ory Kratos exist, but they often come with other tradeoffs


Great. More slop, can't wait.


I am not - I hate AWS(and cloud in general) with a passion - overpriced, you are getting locked in by a closed ecosystem the moment you say "hey this feature is neat it will save me so much work", only to realize that you are stuck paying for it for years if you decide to move away from it. But people are inclined to jump on a hype train and become evangelists for life. Truth is AWS(or GCP or Azure or anything else) is a viable option in two cases:

1. You are making a product with 3 friends on evenings and you want to ship asap without having the capacity to invest and setup infrastructure. 2. You are a huge corporation with tens of thousands of employees and hardware needs that you simply cannot source yourself easily or sort out the collocation of the hardware.

Everyone else - get a dozen second-hand servers, shove them in a rack in a data center and you will own the hardware and everything associated with it at half the price of what you'd be paying AWS in a year.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: