Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

From my perspective it’s just _confusing_ to work in AI right now. We have some massive models that are doing some really neat stuff, and apparently hundreds of millions of people are using them—but I keep wondering: to do _what_, exactly? I’m not asking what the models can do, I’m asking what people want the models to do every day, all the time.

I’ve been shown some neat pictures people made that they thought were cool. I don’t know that I need this every day.

I’ve seen examples of “write an email to my boss”. It would take me longer to explain to ChatGPT what I want than to write these myself.

I’ve seen “write a snippet of code” demos. But I hardly care about this compared to designing a good API; or designing software that is testable, extensible, maintainable, and follows reasonable design principles.

In fact, no one in my extended sphere of friends and family has asked me anything about chatGPT, midjourney, or any of these other models. The only people I hear about these models from are other tech people.

I can see that these models are significantly better than anything before, but I can’t see yet the “killer app”. (For comparison, I don’t remember anyone in my orbit predicting search or social networking being killer apps for the internet—but we all expected things like TV and retail sales to book online.)

What am I missing?



You’re asking “what’s so big about GUIs? Literally nobody has asked to move a pointer around a screen”.

It’s the use cases these thing enable that are important.

Today, I wrote a draft product announcement. Only after I was done did I realize I had written it in a really impersonal third person (“users will be able to”). No big deal, but maybe 10-20 minutes of work to make it energetic and second person (“now you can…”).

30 seconds with chatgpt. “Rewrite with more energy, in the second person, using best practices for announcements”).

Six months ago I would never have asked for that. Today it was glorious and let me move on to focus on more important things.


I mean Google revolutionized search. Apple revolutionized personal computing.

OpenAI revolutionized… rewriting things with slightly different wording?

I’ve seen so many breathless people posting “this would have taken me so long to search” and then I type 3 keywords from their massive prompt they crafted and find it instantly on Google. We’re talking 1000x or more faster. I feel like the same is happening in your comment. How often have I thought “damn I wish I wrote this blog post ever so slightly differently” in my life? Maybe a handful of times? And yes I’m including all generalizations of that question.

But certainly fake girlfriends and summarization will be mid size fields. Image generation has some mid size potential. But these will be spread between many companies.

I really think it has uses no doubt, but is it a revolution? Where? It’s not creative in the valuable sense - media, art, fashion, etc all will adopt it marginally but ultimately it will actually only serve to further the desire for genuine human experience, and cohesive creativity that we see it really falls flat at. It saves some marginal time perhaps if you’re ok sounding like a robot.

Taking into account the downsides it looks like a hype bubble right now to me, and a draw in the long run. There’s just a whole lot of tech people trying to cash in on the hype.


You can program GPT in English.

Let me repeat that: You can program GPT in English. ENGLISH!

You're complaining about the first nuclear test bomb being impractical and uninteresting. How will this change the world? That huge monstrosity had to be affixed to the top of a test gantry and took years of effort by a veritable army of the best and brightest to make! No way it could change war, or geopolitics, or anything. No way..

This is the day after Trinity. The bomb has gone off. A lot of physicists are very excited, some are terrified, and the military is salivating. The politicians are confused and scared, and the general public doesn't even know yet.

That doesn't mean the world hasn't changed, forever.


> You can program GPT in English.

> Let me repeat that: You can program GPT in English. ENGLISH!

How?

Let me repeat that: How?

I had a little script that from time to time parses a list of jobs from a specific board, extracts some categories, inserts them into an SQLite and have a frontend that displays them to me in a way I want.

The board has since changed some things which would mean maybe 2 hours of commitment from me to update the script.

How do I program GPT in English. ENGLISH! To do that for me? What are the steps involved? I've been using ChatGPT and GPT-4 for awhile and I can't imagine what the steps are to make this happen without a lot of back and forth. I can't imagine how to program the infrastructure. I can't imagine how the API endpoint is more than a fancy autocomplete. I need help understanding what it means that I can program it in ENGLISH! (I can also program it in my country's language for what it's worth).

> That doesn't mean the world hasn't changed, forever.

I sort of agree with this.


> make this happen without a lot of back and forth

Perhaps this is the part you're missing. When I've watched people program with ChatGPT it _is_ a lot of back and forth because an enormous amount of context is able to be stored and back referenced. I.e. one wouldn't say "make me a Flappy Bird clone for iOS", they'd start with:

"Give me the code for a starter SpriteKit project". Then

"Now draw a sprite from bird.png and place it in the center of the screen".

"Now make it so the bird sprite will fall as if it's affected by gravity"

I won't bore anyone with how might one go from that all the way to a simple game, but I'm sure you see the idea. There are obviously _huge_ limitations to this approach and professionals will get hit them fast, but the proof is in the pudding: people who can barely code are producing real software through this approach. It's happening.


> Perhaps this is the part you're missing. When I've watched people program with ChatGPT it _is_ a lot of back and forth because an enormous amount of context is able to be stored and back referenced.

I've tried to build a lot of fun stuff with it so far. Haven't been able to properly 'program it in English' for anything non-trivial. Back and forth ended up in loops of not what I wanted. I'm just utterly confused at the difference in experiences I've had with it vs. what some people are preaching.

> There are obviously _huge_ limitations to this approach and professionals will get hit them fast, but the proof is in the pudding: people who can barely code are producing real software through this approach. It's happening.

I've had 4 product people I know try to create products using ChatGPT. All 4 of them basically got stuck on the first steps of whatever they were trying to do. "Where do I have to put this code?", "How do I put it online?", "How do I store user data?", "Where do I get a database from?". Basic questions to any professional, but to them it was impossible to overcome the obstacles from code to deployment.

I don't doubt that it's happening and it will become better in the future; I'm just having a hard time trying to grasp where some people are coming from when my experience as a professional, using it, has been mixed.


i've observed this schism between people who can get LLMs to produce useful output and people who are baffled, I think it's a mixture of two things:

expectations: using to the LLM to break problems into steps, suggest alternatives, using the LLM to help them think through the problem. I think this is the people using it to write emails - myself included, having a loop to dial in the letter allows me to write the letter without the activation energy needed to stare at a blank page

empathy: people who've spent enough time interacting with an LLM get to know how to boss it around. I think some people are able to put themselves in the LLMs shoes and imagine how to steer the attention into a particular semantic subspace where the model has enough context to say something useful.

GPT4 writes boilerplate python and javascript servers for me in one shot because I ask for precisely what I want and tell it what tools to use - I think because I have dialed in my expectation for what it's capable of and I learned how to ask in precise language, I get to be productive with GPT4's code output. Here's a transcript: https://poe.com/lookaroundyou/1512927999932108


Interesting point about empathy. Sorry I'm abusing the comment system to get back to your comment in the future.


Let me give you a simple example. I had to deal with a desynced subtitle file recently. I described the exact nature of the desync (in terms like "at point X1 the offset is Y1, and at X2 it is Y2") to GPT-4 and asked it to write me a Python script to fix this. It did require a couple tweaks to run, but when it did, it "just worked".


"Automatic Language-Agnostic Subtitle Synchronization"

Link: https://github.com/kaegi/alass

It's basically magic.


Honestly don't think it will be long before gpt can read this comment, then politely ask you for the urls of the job board and your git repo and 2 seconds later you will have a pull request to review


You might find this interesting - https://github.com/Torantulino/Auto-GPT

> Auto-GPT is an experimental open-source application showcasing the capabilities of the GPT-4 language model. This program, driven by GPT-4, autonomously develops and manages businesses to increase net worth. As one of the first examples of GPT-4 running fully autonomously, Auto-GPT pushes the boundaries of what is possible with AI.


Perhaps, but I'm talking about right now. I would love to be able to do this as I have 1000 ideas and no time to try them out.


I have some scrapers built in Scrapy, and from my experimentation with GPT4, I bet you could paste in your scraper code, the html source from the website in question (at least the relevant part), and tell GPT4 to update your scraper and you'd get something that's at least 95% correct within 30 seconds.


The same way people used to write code in the early days: trial and error, and a boatload of cursing!


You can't program GTP in anything if you can't program.

If your prompt is garbage then the output will be garbage and if you don't know how to program you won't even realize the output was garbage.

It's not the language part of programming language that is hard. It's the programming part because it means you have to have a good understanding of what you want. Just like a human programmer won't read your mind an AI programmer won't read your mind either.

But I can already foresee bosses dismissing employees that raise issues (performance, maintainability, scalability, etc., etc.) by saying "Look, the AI can do it. So if it can do it you can do it too.". I foresee this because I have already seen it.


> You can't program GTP in anything if you can't program.

That's why it makes this so interesting - this type of automation impacts our jobs directly. Of course, I'm not sure who would use this in a corporate codebase without legal concerns.


Work in a financial place, we have a contract with open AI so they don't hoover our user input. Normal URL is blocked.


> Let me repeat that: You can program GPT in English. ENGLISH!

The very existence of "prompt engineering", numerous discussions about how to prompt ChatGPT in order to get the result you want, etc. imply that while it may be in English, it still requires similar care and attention to do properly as a programming language does.

Which makes me wonder what the advantage of using English is. A formal language seems like it would be more productive and accurate.


For one, GPT-4 requires far less prompt engineering and generally interprets intent better.

The advantage of using English (natural language that is), the humans around you tend to speak it. I don't naturally speak powershell. Instead I want a script that searches for particular filenames, under a particular size, between a particular date in a directory path I specify. I told GPT I wanted that and in a few seconds it dumped out what I needed. It wrote the script in a formal language, which is then interpreted by the machine in an even more formal manner. Let the code deal with accuracy, and lets let language models argue back and forth with humans on intent.


> The advantage of using English (natural language that is), the humans around you tend to speak it.

This is true, but of limited utility. English is so bad at this sort of thing that even native-speaking humans are constantly misunderstanding each other. Especially when it comes to describing things and giving instructions.

That's why we have more formal languages (even ignoring programming languages) for when we need to speak with precision.


That's the other nice thing about ChatGPT - if you say it something and it misunderstands, you can correct it by saying, "no, actually, what I meant is ...". Which, again, is how people generally do that kind of thing outside of programming. The advantage is that you're still remaining on a much higher level of abstraction.

As far as formal languages... GPT doesn't know Lojban well, presumably because of its very small presence in the training data (and dearth of material in general). But it would be interesting to see how training on that specifically would turn out.


> Which, again, is how people generally do that kind of thing outside of programming.

Yes, and with people, that's insufficient if you really need confidence of understanding.

There's a reason that lawyers speak legalese, doctors speak medicalese, etc. These are highly structured languages to minimize confusion.

Even in less technical interactions, when you need to be sure that you understand what someone else is saying, you are taught to rephrase what they said and tell it back to them for confirmation. And there's still a large margin of error even then.

This is why, whenever I have an important conversation at work, I always send an email to the person telling them what I understood from our exchange. In part to check if I understood correctly, but also so that I have a record of the exchange to cover my ass if things go sideways because we didn't understand each other, but thought we did.


Every text interface will eventually include a half baked implementation of a programming language.


> You can program GPT in English. > Let me repeat that: You can program GPT in English. ENGLISH!

What does that mean to "program GPT"? Do you mean program (software) USING GPT?

I thought we already had COBOL, which is pretty much like English so business people can use it. Same for SQL.

And don't we already have lots of low-code or no-code tools? Why do we need to program with ChatGPT if we already are beyond programming?


Not the person you replied to, but I see it the same way. GPT is an English (and other natural language) compiler.

Not in the sense that you get a computer program out (though you can), but in the sense that it can automate anything without even needing a programming language, compiler, and domain specific UX.

Low code and no-code tools still require thinking like a programmer. You define what you need to do, then implement, then get results. GPT often lets you go directly from spec to results.

If the goal is programming, GPT is nothing special. If the goal is quickly reasoning over very abstract instructions, it’s amazing.

The trick is seeing the new use cases. It really does come back to the GUI revolution: if you want to list files in a directory, the CLI is just as good, maybe better. But GUI makes photoshop possible.

GPT makes it possible to say “summarize the status emails I sent over the past year, with one section per quarter and three bullet points per section”. And the magic is that is the programming.


> What does that mean to "program GPT"? Do you mean program (software) USING GPT?

A sibling comment already explained the second part of the question, but there is something I find more exciting. You can program GPT, as in you can tell it to change its behavior. The innumerable "jail break" prompts are just programs written in English which modify GPT itself. Like macros in lisp I guess. The first time I truly saw this potential was when someone showed me you don't actually have to change the temperature of chatGPT in code, you can just tell it to give low and high temperature answers in the prompt[1]. That's programming the model itself in english.

[1] https://news.ycombinator.com/item?id=34801574


The Ford Nucleon was a 1957 concept car that featured a compact nuclear reactor. Look at how well that prediction aged. It's apt that you mention the Trinity test, since 1950s inflated expectations of the applicability of nuclear everything are exactly where we are now.

Perhaps I could interest you in some Radium Water? It's new and trendy and good for your health.


Wikipedia has a fascinating article:

https://en.wikipedia.org/wiki/Radioactive_quackery


> Let me repeat that: You can program GPT in English. ENGLISH!

For problems that are fine being defined ambiguously. Try to program a database in English, let's see where it goes.


You laugh, but this is why SQL reads kinda-sorta like English. People have tried, and failed.

Meanwhile, if you give Chat GPT your database schema and ask it to write a SQL query for a report, it can do that for you.


If anyone wants to see the output of GPT4 when asked to define the tables and some sample queries for a hackernews clone in sqlite:

https://poe.com/lookaroundyou/1512927999932134


This is a very simple case that doesn’t reflect the complexity of a real project. Like so many attempts before to produce code, using a little effort, it degrades when the complexity level increases even slightly. Once there are more tables, ones that have names which cannot be easily translated from English, it breaks down quickly. These types of tools work ok for brand new projects, but work on existing projects will prove to be exponentially harder or more difficult than it is worth.

Nonetheless, it could prove useful for looking up algorithms, patterns, and generating boilerplate code. However, an important issue is will it generate similar code if queried at a later time? Not likely, which will make it less useful or result in an inconsistent codebase. Maybe you can request a version of the code generation? In-house code generators will generate consistent code, so it will be interesting to see how it is used in real projects.


Here's a more extreme example, using SQL as an API to give model access to game world state to reason about it.

https://gist.github.com/int19h/4f5b98bcb9fab124d308efc19e530...

Note that in this case it isn't even asked to write specific queries for specific tasks - it's just given one high-level task and the schema to work with.

You're right, though, that the effectiveness of this approach depends very much on schema design and things like descriptive table/column names etc (and even then sometimes you have to make it more explicit than a human would need). You really need to design the schema around the capabilities of the model for best results, which makes it that much harder to integrate with legacy stuff. Similarly, not all representations of data work equivalently well - originally, I gave the model direct access to the typed object graph, and it handles that much worse than SQL. So if your legacy software has a data model that is not easy to map to relational, too bad.

On the other hand, GPT-4 is already vastly better at this kind of task than GPT-3.5, so I think we can't assume that this will remain a limitation with larger models.


> You really need to design the schema around the capabilities of the model for best results, which makes it that much harder to integrate with legacy stuff.

This may end up being a feature of some high level frameworks … “compatible with ChatGPT” or “designed to work with xxx LLM”.


It will be very amusing if, eventually, our jobs as software engineers will be crafting bespoke AIs to maximize the efficiency of their use by an LLM.


> and then I type 3 keywords from their massive prompt they crafted and find it instantly on Google.

Seems I and you have different Googles and you still have the one I had pre 2010.

For over a decade now, Google has been including things I never asked about to the point where it would sometimes be easier to find it using Marginalia.

Some say it is just because internet has changed and there is less ham and more spam, but the last few months I have been using Kagi and it proves it is possible to create a better search experience.

And, if Google works for you, fine. Maybe you search other topics, use other keywords or are in another bucket wrt experiments, but from my perspective Google is now the same as its predecessors.


I actually agree Google has gone downhill. Yet for the 8 or so examples I’ve tested where I saw hyped GPT results, every single one google answers, usually in the top snippet, always in the first result.

For politics shopping and some other topics it can be terrible, but I don’t think GPT is good at those either.

I’m actually happy to be proven wrong here. If you have some examples let’s test it out. If it’s a true step function improvement I’d expect it to be easy to source examples.


Haven't used Google in a while but let me try.


I think this is a classic case of us overestimating the immediate impact and underestimating the long term impact.

Right now, they are definitely useful time savers, but they need a lot of handholding. Eventually, someone will figure out how to get hundreds of LLMs supervising teams of millions of LLMs to do some really wild stuff that is currently completely impossible.

You could spin up a giant staff the way we do servers now. There has to be a world changing application of that.


I”m not in ML, so excuse this maybe naive question:

> get hundreds of LLMs supervising teams of millions of LLMs

What does this mean or what can you do with this setup… do you mean running LLMs in parallel?


Yes, that's called 'ensembling'. There is a lot of work being done on this kind of solution. One way in which it could work is that you can use multiple models that have been fine tuned for various problems and then use the answer that returns the highest confidence.


You can also have adverserial generation where models given different expertises and attitude can go back and forth criticizing each others work


Sounds like the 'dead internet' is just around the corner!


Ask the LLM to perform a complex task by splitting it into sub tasks to be performed by other LLM instances, then integrate the results...


Is this something like langchain is working towards?


AutoGPT, BabyAGI


We call that “companies”. We just need to apply what we learned in business school to a different set of workers, slightly deficient workers.


> We just need to apply what we learned in business school

Please don't. You've already ruined enough industries. Let the MBAs do finance and Wall Street and leave them out of the chain of command in organizations that make things.


Every time you go to the store and find that the store is still in business and there is food on the shelf, it is because someone went to business school and knows how to optimize demand estimation, pricing, and logistics.

Yes, some MBAs fuck things up. Just like some CS grads fuck things up. But advocating against the study of business is just as naive as advocating against the study of computer science just because there are some bad CS grads.


> Every time you go to the store and find that the store is still in business and there is food on the shelf, it is because someone went to business school

Are you contending that business were not successful before Wharton started pumping out MBAs?

> But advocating against the study of business is just as naive as advocating against the study of computer science

I didn't say 'don't study business', I said 'stick to finance'. MBAs tend to end up destroying innovation and productivity for short term growth and stats.

Jack Welch showed what a successfully motivated 'business oriented' leader can do to an innovative and productive legacy organization when given complete control over it. The MBAs happen to just do it on a smaller scale.


> advocating against the study of business is just as naive as advocating against the study of computer science just because there are some bad CS grads.

Criticizing garbage MBA programs is not criticizing the study of business. Business schools don't study business. They're a place where people make a lot of money selling theories about business that are useless at best and it many places, quite harmful. Learning about business by going to business school is like learning to kiss by reading books about kissing.


That is a great analogy.

I would say that just as every person is unique so is every company unique. And just as there is plenty of pseudoscience plaguing psychology so are MBAs full of pseudoscience. Two fields that are far too obsessed with generalising their advice. Which is not to say that there aren't any useful ideas in these fields. But the vitriolic reaction above is warranted.


Stores existed before the MBA, but MBAs could be why food prices are up 30% since last year.


Take shelter under my protective wings, O, sweet summer children.


>Eventually, someone will figure out how to get hundreds of LLMs supervising teams of millions of LLMs to do some really wild stuff that is currently completely impossible.

This is an intuitive direction. In fact, it’s so intuitive that it’s a little bit odd that nobody seems to have made proper progress with LLM swarm computation.


I've read about people doing it, I haven't read about people achieving anything particularly interesting with it.

It's early days. There will be a GPT 5 I'm sure, maybe that one will be better at teamwork.


This sounds like that old economics joke that says it's impossible to find $20 on the ground, because if it had been there, someone would have already picked it up.


In particular, it's odd that the greatest software developer in the world (ChatGPT) hasn't made progress with LLM swarm computation.


How is "LLM swarm computation" different that single bigger LLM?


The same reason why you don't let Mr Musk do all the work. He can't.

One LLM is limited, one obvious limitation is its context window. Using a swarm of LLMs that each do a little task can alleviate that.

We do it too and it's called delegation.

Edit: BTW, "swarm" is meaningless with LLMs. It can be the same instance, but prompted differently each time.


> The same reason why you don't let Mr Musk do all the work. He can't.

Better to limit his incompetence to one position.


I beg to differ. Imagine him taking down Twitter, Facebook, Instagram, and all the others in one fell swoop!


Context window is a limitation, but have we actually hit the ceiling wrt scaling that? For GPT, you need O(N^2) VRAM to handle larger context sizes, but that is a "I need more hardware" problem ultimately; as I understand, the reason why they don't go higher is because of economic viability of it, not because it couldn't be done in principle. And there are many interesting hardware developments in the pipeline now that the engineers know exactly what kind of compute they can narrowly optimize for.

So, perhaps, there aren't swarms yet just because there are easier ways to scale for now?


I am sure the context window can go up, maybe into the MB range. But I still see delegation as a necessary part of the solution.

For the same reason one genius human does not suddenly need less support staff, they actually need more.

Edit: and why it isn’t here yet is because it’s new and hard.


It's easy to distribute across many computers which communicate with high latency


LLMs are already running distributed on swarms of computers. A swarm of swarms is just a bigger swarm.

So again, what is the actual difference you are imagining?

Or is it just that distributed X is fashionable?


Rather large parts of your brain are more generalized, but in particular places we have more specialized areas. Now, you looking at it would consider it all the same brain most likely, but if you're looking at it in systems thinking view, it's a small separate brain with a slightly different task than the rest of the brain.

If 80% of the processors in a cluster are running 'general LLM' and 20% are running 'math LLM' are they the same cluster? Could you host the cluster in a different data center? What if you want to test different math LLM modules out with the general intelligence?


I think I would consider them split when the different modules are interchangeable so there is de facto an interface.

In the case of the brain, while certain functional regions are highly specialized I would not consider them "a small separate brain". Functional regions are not sub-organs.


Significantly higher latency than you have within a single datacenter. Think "my GPU working with your GPU".


There are already LLMs hosted across the internet (Folding@Home style) instead of in a single data center.

Just because the swarm infrastructure hosting an LLM has higher latency across certain paths does not make it a swarm of LLMs.


> There are already LLMs hosted across the internet (Folding@Home style)

Interesting, I haven't heard of that. Can you name examples?


I read about Petals (1) some time ago here on HN. There are surely others too, but I don't remember the names.

1. https://github.com/bigscience-workshop/petals


It’s a hype bubble for hundreds of years and saying that doesn’t make chatgpt worth any less. I have definitely been surprised by this and gotta say I’m expecting AGI a lot faster now. Even if literally all it did was predict what the average internet user would write in a certain context, that’s huge, cuz when you integrate all the little advantages of all the weird things one person knows another doesn’t, the collective knowledge is worth more than the sum of the parts. A tool which can tap into the sum total of human knowledge 24/7 and more rapidly than I can propose more questions for it, mainly I’m just excited to play with larger context size models so I can include more code and get big picture ideas about groups of stuff that are too numerous for my feeble meat brain to reason about. 7-9 things in working memory has always been the thing that would make humans inferior to AI in the long run. Even if it’s not that insanely smart (but realize: intelligence is a probabilistic concept and computers are great at multiplying probabilities precisely) if the thing can fit more stuff in memory than us and type faster than us and it doesn’t get tired or overwhelmed and give up (imagine your capability in a world where you had no tiredness and unlimited self discipline) in time it’s inevitable the transformers put us all to shame, and the more complicated the topic, the bigger of a shaming it’ll be, since the more complicated topics have exponentially more relations to reason about. Who’s gonna trust a human doctor to diagnose their stuff if the human brain holds 9 things and the AI holds thousands?


"Who’s gonna trust a human doctor to diagnose their stuff if the human brain holds 9 things and the AI holds thousands?"

The human brain can hold much more than 9 things and even though AI will be used in medicine broadly very soon, I really want the final diagnosis done by a human.

Once true AGI arrieves, I might change my opinion, but that might take a while.


9 things is considered a standard for working memory (kind of like processor registers), for people with ADHD it's even less - 3-5.

Try writing a number from one piece of paper to another. If it's more than 7-9 numbers, you won't do it in one shot, unless you spend extra time memorizing it.


That can be increased quite a bit with practice. But it's also not important. It's just the cache memory -- it isn't the limit of what can be learned and recalled.


It is a limit on what you can reason about without a piece of paper.

I’m proficient at math, but my working memory is around 6, so I cannot add two three digit numbers to each other in my head (unless I see numbers to be added in front of me).


OK, but I and nearly all of my friends can, so we have duelling anecdotes here.


The equivalent for computers would be L1 cache on the CPU which is tiny.


More like cpu registers I would say :)


> AI will be used in medicine broadly very soon

We have been hearing this since forever.

Revolutions do happen but not the way we expect. My anecdotical experience: no one in my team of about 30 people developing SW uses ChatGPT or similar in their day to day. This may change, or not.


AI is being used in medicine already. For example, in diagnostics. Most new diagnostics devices (e.g. CT scan, cardiograms) include AI systems that suggest an interpretation and point towards possible problems that a doctor might occasionally miss.

Granted, currently deployed systems are mostly awful, way behind the state of the art, and therefore mostly useless. Maybe it's because designing medical devices and getting them approved takes so long. Maybe it's because the manufacturers put AI in there for marketing purposes only, while assuming nobody will use the suggestiona anyway. In any case, I strongly expect the trend to continue and these systems to become very useful quite soon.


> Who’s gonna trust a human doctor to diagnose their stuff if the human brain holds 9 things and the AI holds thousands?

I will. As another commenter says, the brain isn't limited to 9 things at all. There's no way that I'll trust the diagnosis of a machine that won't understand me.

If a doctor uses AI to help with research, that would be OK. Just so long as the doctor is actually the one doing the diagnosis and prescribing the treatment.


The difference between your search query and theirs is clearly the level of expertise. Chatgpt has a great use case when you get started on a new subject; even with a very cluncky description it can point you into the core concepts of any field. Instead of reading 10 papers with are somewhat related but not what you are looking for, you can spend 3 minutes writing clumsy prompts and that's about it :)


Exactly. Does anybody remember in the early internet days people laughing at their parents for googling things like "please help with my back pain my doctor sucks".

Well, who's laughing now?


To add to this, using ChatGPT feels great in the moment, because it seems to work so well. For example, asking it for an itinerary while traveling gives you something that looks great.

However, once you actually start using it and see that the "ten minute walk" is actually an hour walk, or that a full third of the attractions it has shepherded you to are permanently closed, you realize that building that itinerary yourself from scratch using Google or TripAdvisor would take you less time than manually double checking everything ChatGPT says.

It's also quite surprising that people still think ChatGPT is capable of logic. Even for a complete layperson, all it takes is asking it to draw someone's family tree as an ASCII chart to see that text prediction only goes so far and there's not enough of a relational concept in there to comprise knowledge. There are many examples of asking it to solve famous puzzles with minor variations where it fails spectacularly.

The marketing behind ChatGPT is genius, but there is only so far you can go before the honeymoon is over and people start to really question what you brought to the table. Aside from that, ChatGPT isn't unique in what it can do, and others (including open source) are catching up fast.

That being said, I'd still use it for something like language learning (and other types of learning), where follow up queries (such as why you'd use one word instead of another, or how to rephrase something to be more polite) unlock a significant amount of value. It can also be useful to write trivial code, though I doubt a serious professional would do this (for several reasons, such as privacy and liability). Ultimately, ChatGPT fits squarely under "tool" and not under "intelligence".

It seems that as of right now, the killer app of ChatGPT is the boost in views you get by putting it in the title of your YouTube video.


As for googling, here are some examples of queries you can try and see how it works:

- summary of all the carbon neutral concrete methods, especially ones that can be done in a small industrial workshop as a prototype

- I have allergies in Thailand, mid-february. What may it be related to?

- list all the companies from Japanese stock exchange that have high debt rate

Those are top of my head, but really anything that is either a super-specific niche, or requires merging a few niches together, Google won't help you with.


How do you deal with it straight up lying? My problem with this whole system is, if I’m asking those questions it’s because I don’t understand the field well enough to answer it myself, which means I can’t pick up on if ChatGPT is lying…


Fair, but not completely true. The Thailand examples gives a detailed reasoning. You can use those building blocks to check. If it says Thailand is a cold country and uses that in its argumentation, it's shaky. You don't have to be an expert climatologist to make this judgement.

It's not just one clean answer and we're done. In my experience it is helpful in breaking the problem down into stuff you can Google.


> In my experience it is helpful in breaking the problem down into stuff you can Google.

Yeah I can see that being useful. I’ve also seen a lot of non-technical people straight up accept whatever comes out of it, so that’s a little worrying. It’s true of Google searches too, of course, but at least a google search gives N results someone can check rather than 1.


They straight up accept until they discover first mistakes :)


Fact check with google.

With the example questions I provided, it would take many hours to do research on the subject. GPT provided initial answers instantly, and then fact checking was easy.

That’s what we did with gpt-3. With plugins you can have gpt fact-check itself.

Also, if you have a system for dedicated knowledge, you can use embeddings - with embeddings gpt has very little room for hallucinations, and it can provide detailed references.


For all of those, you HAVE to be okay with a complete garbage answer. Are you?


A machine generating confident bullshit will be the perfect companion for con-men to partially automate their workflow.

Humans are really gullible for the appearance of confidence. And humans are also very prone to wishful thinking.


> Taking into account the downsides it looks like a hype bubble right now to me, and a draw in the long run. There’s just a whole lot of tech people trying to cash in on the hype.

Techies will realize that they are just giving ideas to O̶p̶e̶n̶AI.com, Microsoft Word, Google Docs and Notion. It is just the same AI bros re-selling their hallucinating snake oil chatbot that are under a new narrative for AI.

There is a reason why the only safe serious use-case of LLMs is summarization of existing text, since everything else it does is untrustworthy and is complete bullshit.

Their so-called 'revolution' is a grift.


English speakers tend to miss a huge use case: Translation.


I do wonder where LLM translator would take us to, considering that Japanese version of Bing Image Creator[0] is still proudly displaying a complete nonsense…

0: https://www.bing.com/create

    作成 *芸術* 開始日
    AI を使用した単語


DeepL.com? You dont need a general purpose LLM to do translation.


GPT-4 is a much better translator than Deepl especially for far apart languages.


A 'use-case' done worse with LLMs especially with reliability. Translation is already done without a hallucinating LLM and can be done offline.

Summarization of existing text is the *only* safe and serious use-case for LLMs.


How is summarization "safe"? The summary might be wrong just as well.

The use-case is anything where occasional bullshit output is an acceptable downside to speeding up. More reliable outputs will enable more use-cases.


And what is a business that is fine with occasional (or frequent) bullshit output? Fake-news and spam.


Every business is fine with some frequency of bullshit output at some level. The question is how often exactly it happens and how much harm the bullshit can cause.


My point was that spam is the perfect use case for this tech. Of course there are other possible use cases, but spam and fake news content creation are the perfect fit. AI will enable one to easily clone the writing style of any publication and insert whatever bullshit content and keep up with the publishing cycle with almost zero workforce.

Want a flat-earther version of New York Times (The New York Flat Times)? Done. Want a just slightly insidiously fascist version of NPR? Done. Want a pro-Nato version of RussiaToday (WestRussiaToday)? Done.

And we already know people share stuff without checking for veracity and reliability first.


Machine translation is already not reliable even without LLMs, so it's not weak point of LLM translation.


GPT-4 translates much better than anything else out there, esp. when it comes to idioms and manner of speech.


Notion going all-in on the "AI" stuff is annoying/concerning to me. Mostly just that I live and die by a personal Notion wiki to keep my life organized, and if they eventually tank their service by investing too many resources into features that don't take off and I have to find a new tool to offload my brain into, I'm gonna be pissed...


I went with Logseq and for the first time in a number of years (actually since OneNote 2016, the last self hosted version) I am actually happy with my tooling again.

It doesn't cover everything OneNote 2016 did, but it does a lot more in other areas and it is progressing nicely.


This looks great! Do they have a (decent) mobile app? Being able to jot stuff down whenever, wherever is a make-or-break feature for me...


The app is already usable, at least on iOS, but for now sync is a bit rough around the edges, i.e., I need to verify it is synced or it will overwrite and I have to fix it using the page history which thankfully exist.


> Taking into account the downsides it looks like a hype bubble right now to me.

100% agree and so glad to see someone else say it. I feel like people are losing their minds every time we go through the same hamster wheel.

To hear first hand, in the article above, the effect this is having on ML engineers breaks my heart.


Llm are a tool that understand intent. It makes super easy to compose api into agent and give them a task https://python.langchain.com/en/latest/modules/agents/agents...

That is just the introduction, showcasing what level of sophistication you get with just Google and Wikipedia as tools

Now imagine task rabbit or fiver as tools. Ai can make things happen in the real worlds.

These llm have limited attention but infinite focus. You can parallelize them, you can have one direct a fleet of other llm, you can have llm checking input and outputs for correctness from the other models and feedback that information to the controlling model so that it can improve the promp to the other as it tries to reach it's goal

And the goal can be far fetching (manufacture fake artsy trinket and import them from China to distribute etsy) or nefarious (produce subtle propaganda in a moltitude of wordpress website, register accounts on Wikipedia, reddit, create a sophisticate network of citations)


> OpenAI revolutionized… rewriting things with slightly different wording?

Yes, if you try hard enough, you can try to cast transformational shifts as trifling.

- e.g. “Barteen, Shockley , and Brittain made a smaller version of the vacuum tube.” (transistors)

- “Scientists discovered that light could carry information, like electrical wires do.” (fiber-optics)

The effects (including the harder to measure cultural shifts) matter more than some uncharitable characterization.

Also, the “it is not X” thinking is the result of present fixation. Such argumentation is, at best, quite narrow. Perhaps applicable in specific defined markets and situations but hardly a good mindset for making sense of how the world is changing. Hence the cliché, “The Stone Age didn’t end because we ran out of stone.”

The psychological undertones in the comment above are probably “people, stop exaggerating”. From one overreaction to another, it seems.


The popularity of ChatGPT revolutionized time. For learning, for many kinds of busywork (it's redefining what is and isn't "busywork), for planning. And most important: we don't know what we have yet because it's still being built. It's a tool. It's not the "capabilities" it's what people get out of it.

You mention blogging from the standpoint of writing it all yourself, and then using a tool to tweak it. That's not the revolutionary part. It's collaborating with the tool to write the post.


I need feauteres from n article description and chargpt can just extract it without any effort.

I don't know any other library I could just use with this task with nearly the same quality besides some regex soup.


You act like writing itself doesn't take time and energy. It has sped up my grant writing 6 fold. Any long-ish form writing that I need to do now happens at warp speed


if we only focus on this part. this function represents that most of content creator don't need to convert or reproduce their content to fit another group of customers.


That's very cool, and right now it's a good idea, but I strongly suspect GPT only looks clever and does a good job in isolation. If everyone starts using it product announcements will start looking very similar, and they'll lose a lot of their impact.

This is definitely the case with cover letters for jib applications. The ones written by GPT appear to be pretty obvious - my guesses could be wrong, but after seeing most applications not having a cover letter for years to most applications having one over the past few months, I suspect GPT is involved, and there's a distinct 'style' that seems to be showing up.

Using GPT could be the 'bootstrap ui' of product announcements. It looks great on its own, but put it next to a bunch of other companies and they all fail to stand out.


Question for you-given two identical candidates: One who does not submit a cover letter and one who submits a cover letter that was clearly written by chatgpt—which candidate would you rather interview?

On one hand the duplicities involved with writing a chatgpt cover letter seem to be concerning in a new hire. On the other hand, it shows resourcefulness and going above the line.

I’m tempted to say I’d prefer the gpt cover letter candidate, simply to talk to them about how they got the idea and how they executed, but I’m curious if you feel the same way.


Right now I'd also have a preference for the GPT candidate because it shows a bit of an interest in new tech. In a year when 50% of candidates subject GPT cover letters I think I'll see it as a sign of laziness or trying to hide poor comms skills. Maybe not though. Time will tell.


I'd probably roundfile the chatGPT one. A cover letter needs to be a personal communication from the candidate, not something machine-generated. The writing style is an important signal.

However, I would have zero issue if they used chatgpt to help compose their CV.


I find it interesting it’s very hard to make it do typos. The type of errors humans make. It can do typos but even those seem weird. If I push it to make errors it either over- or undershoots.

The “correctness” of it is a definite give away IMO.


In fairness, lack of typos in a cover letter has traditionally been interpreted as a sign of diligence!

For more general queries its "house style" tends to be really obvious, with all it's "however, it's important to note" and "ultimately it depends on"s and the tendency to flesh out a one sentence answer to the specific question with two paragraphs or five bullet points of detail at a slight tangent to it...


It can imitate styles of little kids with their phonetic spelling. (Though this is so common in English, French etc but not Italian or Turkish where the spelling is very regular)

I am sure it can at least do very common mistakes like it's vs its or "would of" if prompted right since there's a huge body of that kind of work. Or maybe a human needs to add the finishing touches to make it look more human. :)


That distinct ChatGPT style is mostly the product of their RLHF, so it's what you get by default if you don't ask for something more specific. But it's fairly easy to tweak the prompt to make it use whatever style you want, including more terse, less apologetic etc. Don't forget that "write about X in the style of Y" was one of the first things that GPT models could reliably do, long before chat.


Hmm…no, I’m not doubting the value of the interface. I’m asking what’s the “killer app”.

Maybe I should just ask ChatGPT to explain it to me…


It seems like the use of ChatGPT is something like "microtasks". Little things a given person could do but would rather not and so is able to delegate to an automatic thing whose output they can verify.

It seems like it's potential as of today is increasing or seeming to increase the productivity of a segment of white collar workers in the fashion that email and the web did (or might not have). A lot of researchers might not have need for this and so not understand this appeal of this.


Thanks for your comment, I think this is actually very helpful framing! It at least gives me something more tangible to think about—“micro tasks”.


I, too, have been experimenting a bit to see if/how an LLM might help me. This microtask framing jives with my experience.

One of the best examples so far for me (and it's truly micro) was at grocery store. Friend trying to figure out how big of a rice bag to get and avoid not finishing it before a long trip coming up. She knew she ate a couple of cups dry a week.

"I eat 2 cups dry rice per week. Can I finish a 25lb bag in less than 4 months?" "Yes" (it did show its work).

One shot, perfect response. I know this kind of computational thing is what WolframAlpha was for, but that wasn't nearly as reliable. I know I could figure it out myself, but I'd need to find a reasonable figure for the density of rice and probably do some imperial metric conversions and generally futz around for longer than one would want to stand in front of a pallet of rice bags.


50 cups of rice in a 25 lb bag, 4 months = 16 weeks = 32 cups at 2 cups / week. I think there will be some rice left.

Maybe this is one of those examples where this tool gives a confident and wrong answer even when it shows its homework?


Honestly this sounds more like a place for GPT/Wolfram API than just LLM alone.


Really? You're using ChatGPT to do basic math you could easily do in your head?


If we're characterizing LLMs as a kind of input interface I'd point out that the first GUI was released in 1973 by Xerox and the first commercially successful GUI was released in 1984 with the Mac 128K. It took 11 years for someone to answer this question. Sure things move faster these days, but we're still only a couple of months in.

With blockchains there was also a fundamental technological breakthrough (I'd argue less revolutionary than LLMs). The problem was that everyone jumped the gun and proclaimed the killer app had been discovered too soon: cryptocurrency's incarnations to date have yet to demonstrate much utility apart from being a vehicle for speculation. Nakomoto invented the first distributed blockchain in 2008...


It’s programming. The thing is really great at programming, and everything else is icing on the cake. They already got my $20 a month lol


Anecdotally, I've had more writers (screenwriters and copywriters) tell me they're using ChatGPT than programmers. I think people here might underestimate how big a deal it is in "the real world".


Not really programming but generelly producing text (or maybe generally certain structure like ASTs at some point). Programming is a subset.

I expect that this will move lots of things into the world of text that weren't before.


Bingo. No programmer seems to want to program without it anymore. Sounds like a killer app to me.


Most of the economy involves moving around physical things. Construction, transport, nursing and related health occupations (physiotherapy, home aid, etc.), retail and wholesale logistics. Manufacturing employs another chunk. Services for agriculture, fishing, mining also employ more people than are directly employed in those industries and are mostly to do with machines and equipment. Utilities.

Most of the rest is high-touch. People want interactions with humans for important stuff, not with AI. Remote teaching was an unmitigated disaster for most school students: how will AI teachers do, do you think? Attempts at robot police and security guards haven't gone down very well to date. It'll be a while before there are AI EMTs and firemen.

So there are grounds for skepticism.


Well clearly you’re going to be sitting on the sidelines for the next few years.


I guess if you don’t like writing, this is good for you. However, I like writing. Moreover, many people are lousy at it (not that I couldn’t improve). I’m not sure I want something trained on lots of mediocre writing doing editing for me.


LLMs are trained on language. Not mediocre language. This is why models can be fine-tuned in one language and then see the benefits in other languages. How much longer with this fundamental misunderstanding of these models continue and how often will they be put forth by people who are worried about task X that they enjoy being replaced?


> What am I missing?

I'm going to guess (assume) you probably haven't worked in a 'real' business. A place where elbow grease still does the majority of the work and where Windows 8 was only just phased out.

The killer app (to me) in case of GPT is GPT itself, not ChatGPT. ChatGPT just allows me to easily test use cases for GPT. There are many interesting use cases for those elbow grease businesses for GPT. For example:

Data entry. There's still a lot of data entry being done from unstructured text. Where specifics like names and addresses need to be extracted from letters and emails and contracts and other stuff. I've worked on these challenges before using different strategies and GPT blows my mind with what it can do just by asking to grab this data and format it as JSON. Is it 100% correct? Nope. You still need people to review the data (depending on the use case), but that already saves tons of work.

Categorization. Some companies still get tons of emails that need to be forwarded to specific departments. This is another thing that GPT does surprisingly well out of the box.

And that's just GPT. There are many other legacy business processes that can be automated (partially) by other models that are coming out right now. Even just 'segment anything' that Meta just released is incredibly useful for many use cases I've seen in my daily work.

Killer apps are always a combination of a technology to solve a real world problem. If you don't venture into the real world and only stay part of the tech world, seeing the killer app is very difficult and ends up leading to Juicero-like products.


Right, questioning the value of a chatbot is a good indication someone hasn't had real job. Totally legit characterization that doesn't make the whole thing sound like a confidence game.


> Please respond to the strongest plausible interpretation of what someone says, not a weaker one that's easier to criticize. Assume good faith.

https://news.ycombinator.com/newsguidelines.html

> questioning the value of a chatbot

Original commenter was questioning the value of the entire field.

> hasn’t had a real job”

This is different than “haven't worked in a 'real' business”.


Ironic given yours and others emphasis on my choosing "chatbot" to summarise llms as a reason to dismiss my comment, along with the rest of the pedantry. The upstream post dismissed / insulted the person for questioning value, which was what I called out.

If you really had wanted to get into the "HN rules" game, you could at least have cited "don't be snarky"


> dismissed / insulted the person

That wasn't even remotely my goal and I'm disappointed that my choice of words made it seem like it was. I purposefully added quotes to the word 'real' in my comment since any business is a real business and made it clear it was an assumption, not a fact.

It's just that many tech workers often haven't worked outside of tech and therefore are blind to issues outside of the tech world, like manual data entry, because they assume that must all be automated. It's exactly the same the other way around, people in what I called 'real' businesses are blind to what tech can to to improve business processes because they have no clue about what's available and possible.


Thanks for putting in the effort to clarify. I get it.

Translation: “Real” businesses sort of make their own gravity.


Fair enough, thanks for clarifying.


You weakened the argument by portraying it as more insulting against a narrower claim than it was.

But yes, also snarky.


Real Job vs. outdated job.

You don't want to know how many companies still get paper bills scan them and add them to their system half manually.

And normal people without scripting experience never had the tools to just do a little bit of text analysis without tools like chargpt.


They weren't questioning the value if the chat bot, they were questioning the value of the models based on only the high profile use cases such as a now (in)famous chat bot. That, to me, showed that they have a narrow view of the problems that this stuff could solve when applied outside of those high profile use cases. So I made an (explicit) assumption that they mostly worked in tech and not outside of tech, which limits the view of the world outside of tech.


I don't think donkeyd was trying to insult the person he was replying to, but he definitely could have worded that better. I think he's using "real job" in a derogatory sense directed at businesses, how most of them are so inefficient and rote that they'd stand to gain from the roteness of something like GPT, and that the person he's replying to, perhaps fortunately, hasn't experienced that type of business.


they mean they didn't have a job where neither ai or bad workers should be used.


Data entry is a really interesting one to me. We’ve been replacing a legacy (read > 1 MLoC with no tests), system on and off for a few years. The original system had a ton of double entry or manual data entry and the human error rate is noticeable. If GPT could have automated this with a similar or reduced error rate then we would have considered it a win.

The real win long term remains killing off manual entry any time it’s possible, but GPT offers a nice patch.


> There's still a lot of data entry being done from unstructured text. Where specifics like names and addresses need to be extracted from letters and emails and contracts and other stuff. I've worked on these challenges before using different strategies and GPT blows my mind with what it can do just by asking to grab this data and format it as JSON.

Is this data confidential or something you are willing to send to anyone? If the former, you probably shouldn't be sending it to an AI company that retains the data for its own purposes.


This is very confidential data, which is why the current implementation is run on-premise and of course I'm definitely not using production data to test GPT capabilities.


That's good! There have recently been news of companies leaking trade secrets and confidential data by sharing it with ChatGPT: https://www.techradar.com/news/samsung-workers-leaked-compan...


Well, in the organization I was doing this, they had similar issues before I started, with external consultants doing stupid stuff. Luckily, my team wasn't as stupid as these people and integrity was also very high. Which was very, very necessary considering the data we were working with.


Nice call about "Categorization". I would put a lot of "Data Entry" in the same category. Excellent ideas!


Latvian is a language, that's spoken by less than 2 million people. It's irrelevant outside of Latvia.

All of our government funded researchers who worked on natural language processing can now throw their work in the trash and resign. ChatGPT is leagues better than anything that they've done. And OpenAI weren't even trying.

Windows wasn't even localised in Latvian properly until very recently. Google translate still spits out ridiculous translations. (Though it's better than before). Most software isn't even available in Latvian. Almost no video games are in Latvian. Only the most popular books and movies are being translated. Interested in something less popular - you better learn other languages.

And now ChatGPT comes out and I can ask it to write C++ functions in Latvian. I don't need to learn English to be a programmer any more. Nothing like that has been done before. It translates stuff way better than google. And it will only get better.

Imagine that there's a book that only 50 people from Latvia are interested in. Human translators aren't going to bother. But ChatGPT can do it easily.

This is a very big deal.


Why is Latvian NLP research in trash?

Real researchers build on advances not quit because of them. I'm sure GPT is not as optimised as it could be to process non-English text. There's clearly a lot of work to do. At least this is true for south asian languages and I'm sure is true even for popular Western languages like French or German


Most pre-GPT NLP research is obsolete for all languages, but this was understood before ChatGPT.


This also works in reverse, BTW - think of all the obscure books, movies etc that were never translated because it wasn't worth the effort.


Quality machine translation for less common languages would be a service worth paying for. Any links to examples and evaluation of ChatGPT capabilities?



It's a hype bubble. There's a small group if die-hards in the tech-adjacent VC community that talk it up to no end and attack anyone who disagrees (I just saw a guy ask you if you were "virtue signaling" with your post - lol. And there's lots of people exploring how it could potentially be commercially useful - just like with blockchain. As everyone points out, it's got more going for it than blockchain did because it does something more tangible.

But the jury's still out on the long-run commercial potential of pretty good autocomplete or chat-as-search. It's probably more than zero (like blockchain), but it won't "change everything".

I'd guess that 90% people either agree with your observation or don't notice / care. It's just that these things bring out the vocal defenders, usually a sign people know deep down it's a bit of BS and post out of insecurity


I've got access to GPT-4 at work. There have been many times that I'll encounter a bug while programming, I'll paste my code to GPT, tell it what my code is currently doing, and tell it what I actually want the code to do. It then explains what I'm doing wrong and what I can do to fix it. I've had success with this over 90% of the time. This saves me a significant amount of time that would have otherwise been spent hunting down solutions on Google, Stack Overflow, GitHub issues, etc.

I don't know what else to say other than that I would not willingly go back to life without GPT. The value speaks for itself to me.


How simple are your bugs? The bugs I usually have to fix at work involve edge cases around complex user interactions that are less programming bugs and more "well we didn't really think about this particular user interaction when developing this feature" - things that are usually simple 1 liner fixes but can take hours to figure out based on back and forth with product to figure out if it's even a bug and tracing where exactly the data/interaction in the code comes from.

Until I can paste in my entire codebase and the entire history of the product development process into GPT I don't see how it can help.

The bugs that it easily fixes, are generally the bugs whose errors i can copy/paste into google and find an immediate answer on stackoverflow


How is this possible? What do you guys do at work? I haven't had success with neither GPT-4 (did you build your own API calling tool for it? Do you just paste it in their Playground?) nor with GitHub Copilot in actually delivering anywhere close to 90% of the time. It usually misses a whole lot of context.

It feels like it would work for perfectly encapsulated small single purpose functions, which of course sounds great but in reality not many projects are structured like this.


I'm doing frontend work (React/TypeScript), so that probably helps since I am working with relatively small components.


I've used it to try and generate some rather small components in React / TypeScript myself, and what it did to arrays of refs with hook calls inside useState hooks initialization function, and the fact that I couldn't get it to fix its issues by doing what people suggest ("just copy paste the error"), or by trying to reason with it, made me not trust it so much. The output code is also pretty low quality in my experience and opinion.


I think you're wrong. I never bought into the blockchain hype, and blockchain added literally nothing to my life other than some speculation right at the beginning. I would pay $200 a month for ChatGPT, today, without thinking twice about it. That alone makes me think it's a much, much bigger deal than you think it is.


Why is this downvoted? It is a good point. My flatmate has a PhD in comp sci, does AI/ML research. He is now paying for ChatGPT "pro" because he is non-native in English and needs a bit of help to improve when writing his papers. He said ChatGPT is so good that he is learning a lot about English by just reading the corrections/improvements. And yes, we have talked/agreed about the scam/sham of blockchain/defi.


> it does something more tangible

it does something more superficially apparent to naive people. decentralization is extremely more important. it will be one of the main ingredients of direct democracy.


I thought about learning Go and Fiber to build the backend of a side project of mine and I did but as it goes for any new language / stack I wasn't feeling confident in it. Then ChatGPT came out and I thought what the hell, let's see what all the fuss is about.

So I asked it to write me a struct for a table with the "id, name, longitude, latitude, news" columns. That worked well, I was surprised it automatically inferred the data types for said columns.

Then I asked it to write endpoints for retrieving a record from that table and it did so perfectly which again I was surprised by. Asked it to add endpoints for adding records and retrieving all records. Again, no bugs, perfect code. At the end I asked it to create a python script to test the API and it did so flawlessly.

Next day I created a docker env with postgres and went to test the code but it didn't work, turns out it wrote it with mysql in mind so went back and told it to rewrite the entire code with postgres in mind and again it did so flawlessly so overall writing this small API endpoint took maybe 30-60 min.

Considering I was a total newbie at Go this probably would have taken me several hours to complete successfully and this code is basically just boilerplate. I don't care to learn it by heart so I can be more productive in the future. Now that I have ChatGPT I basically don't have to, I don't have to write python to speed up my dev time, I can just have ChatGPT write the basic stuff in a highly performant language. It removed the only drawback which was more boilerplate.


> Again, no bugs, perfect code

How do you know? Either you know enough Go to be able to tell, and thus you'd be able to write this yourself, or you don't and thus you can't really judge.

I mean, it probably is right, but this "I don't know something, but I trust what the chatbot told me" is what worries me about the rise of the LLMs.


Reading and validating code is way easier than figuring out how to write it in the first place. A lot of programming is the same - functions, conditionals, loops, libraries, etc.. The hard part is mostly semantics across languages. Way easier for GPT to do the first pass, then go through it and read up on the parts you don't understand.


It isn't in my experience. Understanding code that you did not write for a problem that you do not fully grasp is a lot of work. It's hard enough when you did write the code and when you do fully grasp the problem, which you'd have to if you were to write it yourself. Typically that's what first draft code is for: to see if you actually understand the problem.


Do you find it takes more time to review a PR than it took for someone to figure out the solution and implement it?


Understanding the code of a fully fledged application is hard.

But with ChatGPT you can build it piece by piece.


With a brain you can also build it piece by piece, in fact I don't know of any other way of writing a large software system than doing it piece by piece.


Sure, but the argument was that reading an entire application's code is hard, therefore GPT-4 is counterproductive.


Validating code you understand, sure. Validating code for a language you don't know, I don't see how you could.

E.g. say you speak python, how would you know how to cleanup memory in C, since you never had to do it? Would you even know that you have to?


> Reading and validating code is way easier than figuring out how to write it in the first place

Not at all. I would even say it is harder because the code being reviewed is priming you.

All code is a leaky abstraction and if you don't know what to look for you just won't see it.


That's just not true. I expect software to become a magnitude shittier soon because of this attitude. After that I expect software to become unbelievably good, because thanks to AI we will have the ability to prove correctness of software much more cheaply than before, and to design on a higher level than before. After all, you don't want the AI to help you generate boilerplate code, you want the AI to help you avoid boilerplate code.


It is true if you can understand the code it writes. Maybe you are worried junior programmers are gonna churn out code using ChatGPT but to be honest I'd rather trust code from ChatGPT than from a junior, I feel ChatGPT writes better code on average.


I'd argue you only THINK you understand the code. If it generates a 1000 lines, which look like they are doing the right thing, will you be diligent enough to go through every single one of them? This literally can only work for extremely boilerplate code (which is maybe the code most of programmers write), but for most code I write I need a mental model of it, and constructing that model myself is easier than to try to learn it from some code. Of course ChatGPT can work as an inspiration, especially for working with unfamiliar APIs.


Because what the code does is simple enough, connects to a database, runs some queries and returns the results. I know it worked because I tested it and for that kind of code there aren't really edge cases. Even if I can't validate or catch syntax errors because I'm not yet experienced enough I can see the overall structure of the code and what it does and if it runs then there's no syntax errors.

It's like with riddles, someone asks you a riddle you think and you think and you draw a blank but if you are given the answer you can instantly validate it even if you didn't know the answer before hand, same with this.


On one hand one may not know the exact syntax for a function call but still understand looking at code where a function is defined and where it's called, those are two different knowledge set, the first doesn't transfer from other languages but the second does.

On the other hand you can ask gpt to write test and validate at the outside layer that data is transformed the way you need to.


> I mean, it probably is right, but this "I don't know something, but I trust what the chatbot told me" is what worries me about the rise of the LLMs.

I share the same worry with regards to the way humans will use AI, along with worries about enabling various antisocial behaviours at scale.


Stop looking for the “killer app”; that metaphor is not useful here. Language models don’t need to have one, two, or a hundred “killer apps”. They are likely going to be highly distributed. They can revolutionize every textual interaction point with any person, group, or organization. For worse and better.

Instead, pay attention to how industries and people redefine themselves. There are going to be winners and losers.


I found ChatGPT helpful in a bunch of diverse situations. I'll go into detail below, but overall I think the most valuable thing it can do is understand questions that takes both knowledge (that I don't have and want to acquire) and understanding (that prior AI technologies did not have) and provide meaningful answers. I suppose I find it very useful for my own education - with the understanding that I should use 2 or 3 grains of salt when reading its answers.

For example, I am studying Japanese and often encounter expressions that seems to mean the same thing; I would ask my teacher, but her time is limited. I can instead ask ChatGPT and only bring to my teacher the questions whose ChatGPT answers did not convince me.

Another example: I like understanding why there are certain steps in a recipe. It would be hard to find someone with the knowledge and time to answer the question, never mind I should pay for their time. ChatGPT can explain what I want to know at the level of detail that I want.

I was also able to get a decent understanding of a mathematical question I had no business understanding by recursively asking questions until I was able to link its answers to my own knowledge.

It was also able to answer questions about the Spring framework that I had while reading the documentation itself. In that context, going in rabbit holes severely slows down learning and has the potential to just get me lost.


Be wary with ChatGPT and Japanese. As someone fluent enough and using ChatGPT for inspiration in writing, as well as to find words that either I have "on the tip of my tongue" but fail to remember, or find idiomatic expressions that I may or may not have heard of, ChatGPT can come up with weird hallucinations (Edit: to the point of making up words, I guess mainly because its tokens are at best at the character level ; knowing that, it's amazing it performs as well as it does, honestly).

I always double check with Google searches, but at least ChatGPT gets me somewhere where I can actually search for something useful.

Relatedly, sometimes, even with an initial prompt I've used in the past to make it do what I need to a text in Japanese, and despite everything I type being in Japanese, sometimes it decides to respond entirely in English.

It can also be really bad with context. Because in Japanese, the subject is often omitted because it's known from context, ChatGPT often mixes things up when rewording or summarizing.


Are you using GPT4 or 3?


GPT-3. I assumed that's what most people mean with ChatGPT. Usually, people specify for 3.5 or 4.

Is GPT-4 significantly better with Japanese?


GPT-4 is dramatically better with everything except speed. The minutia most people in this thread are complaining about is almost completely solved by GPT-4.

For example I used it over like 30 minutes to conceptualize, solve, and write some code that draws a graphic for a simple physics problem (0-shot) that I could intuitively understand but had no (physics and math knowledge) tools to calculate properly and it was a great experience.

It let me pick and change how what properties i wanted represented on the graph, knew what center of mass means, how to calculate it for a weird object, got something that feels correct to me, drew it out with various representations where i used color, size, shape to represent the various distances, weights, clusters, etc.

My non programmer friends have used generative networks to make designs for prints, incredible and generally accurate folk art, full mobile games.

People are sleeping on this. Don't be them.


Specific examples don't generalize though. My experience with ChatGPT (GPT-3) is that it is vastly better at dealing with English than Japanese. That is still probably true with GPT-4, but that doesn't mean there hasn't been progress in GPT-4 with Japanese, which a sibling comment says is the case. But it's not because it does better at $task that you can extrapolate to a different one. People have reported GPT-3.5 being better at some things, after all.


I cannot speak for Japanese, but I'd say that GPT-4 is better at Russian.

That the model is better in English is no surprise given that most of its training corpus is in English. In fact, based on the sentence structure of the output when it speaks in Russian, it's clear that what's happening there is some kind of real-time translation from English.

That aside, I have yet to see any task on which GPT-4 wasn't at least as good as, or better than, GPT-3.5. I'd love to experiment with that. Do you recall any specific examples?


Yeah that's fair, it may be that Japanese ended up getting pruned out. I sincerely doubt that though, it seems very, very unlikely.


It's night and day in Japanese.


I guess I should give it a spin then.


GPT-4 still unfortunately makes up words. Which is not totally surprising considering its tokenizer.


ChatGPT is 3.5, and the one mostly used is 3.5-turbo, that's the only free one and it's the default if you pay (though you can pick the legacy 3.5, or 4).


A lot of your examples you can just google search for the answer - and at least google won’t hallucinate; worst case the websites are wrong but GPT’s training data can be wrong too.


You are not missing anything. It's a bubble. This is tech 101 and all over again. You gotta hype some new thing whether it's Web 2.0, Crypto or AI.

LLM have proven to be great as a gimmick or making rather okay localized approximations. You can create a good enough image without any designing skills, or have auto-completion on steroids. However, there is no proof that this same tech can extrapolate to the next level.

Most startups are putting their eggs on this single basket. I have the feeling that this will be what triggers another AI winter and some VCs will holding the bags... but why do I care?


You couldn’t be further from the truth.

This is the single greatest leap in productivity we’ve had in the last 100 years.

Last week, I used GPT-4 to write my code. Later, I used it to analyze 100+ websites and come up with a personalized pitch for a relevant plugin/product idea - something that would have taken me 100+ hours.


If it makes you feel better (or worse?). I am using it daily and extensively. It is a productivity boost but not something that will propel me to god-level and the more I use it (Codex in particular) the more I discover its limitations; and that's where my conclusion has came from: It is too limited beyond the local scope it has. It is not clear (and there are no indications) it can make a jump from that local scope.


It seems then that it's usefulness depends on the intersection of what you want/need to do, what the model is capable of doing, and what you're capable of getting the model to do.

In some cases the impact might be enormous, and in others perhaps less so. One thing is for sure, the models are getting more capable, and along with that people are investing time/effort/money into improving their capability to leverage what the models can do.


Are you using GPT-3.5 or GPT-4?


How did you provide the info for those 100 websites to GPT?


How did you make it analyse 100+ websites? Scraping html and pasting it to gpt, or some other way?

I'm doing similar things, and wondering how other people handle it.


How did you use it to analyse websites?

Did it go scrape stuff for you or did already have the data or did you paste in the website data?


It has plugins now. It can search the web. It's connected to Wolfram. Zapier. It can do things.

Even before that you could hack something together. I told it how to request an image be included in responses - it includes [fluffy unicorn], I parse that out, feed it to another GPT to get a better description, then feed that to DALL-E to get the image to include


Yes, but it hallucinates sometimes. /s


I mean, that’s exactly what it’s being used for — hallucination.


Can you elaborate on this? Or are you just being cheeky?


When you use it for brainstorming, hallucination is not that big of a deal - you need to verify ideas anyway. And it's great at "thinking" outside of the box, if you prompt it right.


Ironically you've proven you have no value add over GPT-4 by choosing to use it to complete your tasks.


That seems like it supports his point pretty well. If he can use it to basically replace most of his work, then it doesn't seem like a bubble at all


What it does is it makes you multidisciplinary. You might be a coder who struggles with writing social media posts and blog headlines. GPT fills that shortcoming.

It’s also a massive skill multiplier. It can turn someone with a 3-6 months knowledge of a discipline equivalent to someone with 2-3 years (or even more) in the field.


1000x this.


I believe that you are half right. AI is going to radically change the world, but not now, maybe in 10 years. The pace of research is slower than the expectations of investors. This will lead to the perception of a boom and bust and AI winter, but in reality there will be slow steady improvements. It’s just been commercialized too soon.


There is also the other (very high in my opinion) possibility that LLM will hit a wall in regards to performance and a new model is required. It's unfortunate that the current VC landscape encourages this single shot mentality instead of diversification. This will not help in the AI winter since most people will be invested in the same thing.


Wouldn't it stand to reason that LLMs should be obsolete in about a year due to the singularity event and the rapid speed which we're moving towards AGI and then onto exponential technology growth due to ASI?

At this point, as an AI researcher, I guess you'd just have to sit back and watch is all unfold, very soon everything you do is obsolete almost immediately.


This suggests that LLM is the path to take us to AGI. There is no proof for that. That seems to be the current bet, but my thinking (and it is strictly mine) that we are still technologically under-powered to achieve such a leap. Maybe in 10 years, or in one year or in a hundred of years. However, for that leap (and strictly my opinion), we need significant infrastructure leaps. Something like processors are x100 faster, or your laptop is powered by a 5Ghz 200 cores CPU...

As it stands, we can't even get the Mark Metaverse right. You are trying to convince me that we have the infra. for AGI? Not convinced.


Three observations here. Firstly it has been a really eye opening experience watching the innovation around Stable Diffusion and locally run LLMs and seeing that the unoptimized research code that needed such beefy hardware could actually be optimized to run on consumer hardware given sufficient motivation.

Secondly it wasn't obvious that deep learning was going to work as well as it did if you simply threw enough compute at it. Now that this tech has reached critical mass there is a tonne more money being poured into infra to support it.

Lastly, compute power is increasing as always. Nvidia releasing H100 and also their recent work on computational lithography. Also DeepMind finding new state-of-the-art algorithms for doing matrix multiplication with AlphaTensor. You can kinda already see the positive feedback loop in action.

I dunno... at this point I just wouldn't bet against the trajectory that we're on.


What actually is the trajectory we're on and what will we do once there?


So, it seems the trajectory is one of increasing generality and capability of models and increasing reliance on them.

If it's at all possible to improve our technology then we will. If we improve it it increases in utility. If it increases in utility we use it more.

What other thesis is there?


The model architecture has stayed roughly the same since the original AIAYN transformer in 2017. That’s 6 years of nothing fundamental happening.

Now, obviously the models have got hugely better in capabilities since BERT. Everything else has advanced. Tweaking, tuning and scaling have delivered true intelligence, albeit sub-human. But it seems unlikely that transformers are what take us to human-parity AGI and beyond, because the more we optimize these word predictors the more we find their limitations.

The lack of architecture changes over the last 6 years creates a huge amount of “potential energy”. A new model architecture might well push us over the human-parity threshold. It wouldn’t surprise me if I wake up one day to find that transformers are obsolete and Google has trained a human-parity AGI with a new arch.

This could happen tomorrow or in 20 years, transformers had an easy discovery path from RNNs, to RNNs with attention mechanisms, to Transformers. Architecture X seems to have a much more obscure discovery path.


It's certainly going to be very interesting what comes out of the training runs that are going to be done on giant clusters of H100s.


I also believe we might see something unfold but we still don't know.

It would be wrong though not to keep an very close eye on it or also to embrace it because if it will not just happen in the next 20 years you still need to earn money and with expertise in ml you might be better of


All current indications says otherwise.

Web3 or crypto didn't disrupt markets.

Chatgpt or let's say ml already did and there is no writing on the wall it will stop. Contrary it shows how much potential there is and how excited a lot of people are.

We have already ml now in office, in bing, in Google workspace. There is midjourney, SD etc.

You can find ml generated porn pictures.

We have constantly news about advantages in multiple ml fields.

This is so different to crypto and stuff.


- Amazing translations (this alone is a game changer)

- Language learning... The ability to improvise realistic conversations is huge. I can ask to talk about cooking a specific dish or a sport!

- As others have noted, refining documents similar to grammarly

- Looking for a product with extremely specific features (though it isn't very good at comparing yet)

- Searches that are too vague and complicated to articulate to a search engine or use exact matching ...


I don’t think people dismissing AI have access to GPT-4 yet. GPT-3.5 did feel like a gimmick.

GPT-4 feels like a team of really capable human interns.

The potential really goes wild once you connect it to the internet and use stuff like autoGPT


This makes a ton of sense.

I've been struggling to reconcile my personal experience with what I'm reading - it was so strange reading such dismissive comments by such a knowledgeable community about a new technology that's such an obvious game changer.

GPT-3.5 was easy to dismiss, but GPT-4 is incredible.


Didn't we have amazing translations for a long time?

I do not notice a significant improvement between new models and last year's deepl/Google translate.


Sounds like you're a native english speaker? Translations to english often worked "well enough" with google. But as soon as you tried the other way arround, or from one non english language to another, the results where often fully incomprehensive. Btw. the same goes for t2s and s2t. It only got slightly usable in the last few years, therefore wasn't really adopted in many non english countries. As you can see, there is a huge market opening up. Then take voice synthesis into the picture. I think the simple amount of recent changes, including LLMs, will steamroll a totally new media envoirenment and with it substantial social and economical changes.

For me its less the capabilities of LLMs, but the speed and the inevitability of change. You can choose to ignore it, but you soon will be outdated then. Just like if someone would try to work an office job without using digital machines. Maybe you can still do it, but who would hire you?


No, I'm an Italian native speaker mostly translating to Hungarian :), it's less good than English but it seems to me it works well enough with both.

First time I tried chatgpt with Hungarian it actually refused to work, so I'm not sure it can be that much better.


Language pairing is important. Machine translation between English and German/Nederlands has been excellent for years. English and Korean/Japanese: Awful for years. On the last couple, did it get much better. I am sure ChatGPT12 will have virtually native level translation. Maybe it will be integrated into Kindle so you that can buy books in 25 diff languages, then get a ChatGPT-translated version with one button click.


DeepL is really good, better than Google Trandlate. ChatGPT is about the same, but it can also explain the translation and the grammar. It might sometimes be wrong but it’s usually good enough at explaining things to point in the right direction.


For translations, I use Google translate. Should I switch?


Yo, check this out! DeepL is like way better than Google Translate, so you should totally use it. But, GPT-4 is like a game-changer, dude! It's obvi way better than DeepL, and it can do all sorts of stuff like rewording, explaining in detail, and changing up the style. This text was translated from Japanese to gamer-style English by ChatGPT-4, no joke!


This passes for “gamer-style” English?


For something machine translated from Japanese? Yes, it's astonishing, actually.


I had to read GP twice, plus your comment to realise it might been generated by a GPT-ish service. Damn, this stuff is getting good.


LLMs are better at context - as far as I can tell Google Translate only does one sentence at a time.

With GPT-3 it infers things like gender and formality from earlier context.

Not clear yet how good GPT-4 is as OpenAI won’t say what training data it has, even rough volumes, in each human language.

Needs some thorough research testing it.


DeepL [1] usually does better in my experience, but I suspect Google Translate has gotten better since last time I used it.

[1] www.deepl.com


I’ll give you a concrete example from a few days ago in my job.

I needed a quick utility window in the Unity editor to see what animations could fire animation events and what those were.

I’m somewhat familiar with the editor API, enough to know what to Google and roughly where to go in the docs. I don’t do it enough though to really learn it beyond that point. So I’d estimate I could spend maybe one and a half hours, counting research, coding something, testing it and then context-switching back to what I was working on.

On a whim I asked ChatGPT (GPT-4) if it could do it for me. Formulating the prompt took a few minutes. I included a short bullet list of what I wanted and told it what Unity version I was on.

In almost an instant, it did it. I copied the code into a new file, added it to my project and it worked.

Time from idea to the first working version was around 10 minutes.

I asked for some minor refinement and then asked how I could extend it. It gave me starting points and taught me something new about Unity. All that slow doc searching, Google searching and forum-trawling was gone.

It’s like having my own personal dev assistant.


One worry I have for the long term is: how will we learn or adopt new programming languages in the future, when LLM du-jour knows nothing about Language X? Will we be stuck with what we’ve got because LLMs make us too productive in them?


I'm very worried that machine learning will lead to stagnation in general, in programming and in art. People will accept the machine output and call it a day instead of making something new.


I thought about this as well.

As long as the new language is well-documented, it should be simple to teach the LLM to use it.


But for a new language, there isn’t a large body of work to train the LLM on what idiomatic code looks like. It really worries me, that we’re going to be stuck at this local maximum


> I’ve seen examples of “write an email to my boss”. It would take me longer to explain to ChatGPT what I want than to write these myself.

You can integrate ChatGPT here to help with the proofreading and editing. If you have a list of points, you have have ChatGPT write an email, then integrate its changes. This is useful especially if English is not your first language. ([append]) Here's a quick example. These emails aren't great, but might be better than what I can come up with myself in 2 minutes. https://pastebin.com/dD22gR4y

> I’ve seen “write a snippet of code” demos. But I hardly care about this compared to designing a good API; or designing software that is testable, extensible, maintainable, and follows reasonable design principles.

It's super helpful when starting in a new space. I needed to write a python data munging script the other day. Using a few ChatGPT queries I found dependencies and understood the basics of using them. I still had to check the docs, but I jumped past the "tutorial" and straight to "API reference".

> In fact, no one in my extended sphere of friends and family has asked me anything about chatGPT, midjourney, or any of these other models. The only people I hear about these models from are other tech people.

Counter-anecdote, I met two copywriters working for media publications (non-tech) who both have had GPT-based services integrated into their workflows by the company management.


>I’ve seen “write a snippet of code” demos.

I've used this part of ChatGPT before. Incredibly useful for getting the syntax of some library that you're going to use once in your life, then never again.

Had a sysadmin mate do something similar to generate a simple Chrome plugin for internal use at his work.

That alone justifies the price for ChatGPT Plus, IMO.


Agreed. It’s also great for things like “zip up all the files in this folder, upload the zip file to s3, and output a progress bar while uploading”. It’s not difficult in an architectural or algorithmic sense, but it takes time to refresh on the APIs, find a progress bar library, and how do presigned s3 posts work again? And you can make little mistakes, forget error handling, etc. You could easily spend an hour or more on something like this, especially if you haven’t done it before in x language.

With ChatGPT it takes a few minutes. It can add up quite dramatically when you have a bunch of these kinds of tasks on the todo list. It does feel revolutionary to me as a productivity enhancer.


I used it to knock up a browser extension tonight. My knowledge of JS is basically non-existent. It definitely paid for itself this month.


its good for so many things:

Need to write a speech as a best man

Write a eulogy for a deceased family member

Recently used it to speed up the time to fix an oauth problem in code I didnt write, knowing nothing about oauth. Replaces stackexchange with much more tailored answers.

can take hundreds of pages of text and distill it down to an executive summary of any length

Daughter uses it to explain how to solve algebra problems, not just give an answer. Will completely change education.

Marketing using it for all content. Devs hate writing blog posts, it will take some code as an input and write a blog post about what it does. Can use it to come up with questions to seed a podcast

Paralegals are virtually redundant. AI with all legal caselaw making it much easier

HR using it to write customized offer letters, review resumes, etc.

It isnt just about asking a question and getting an answer. You can keep adding context and the answers keep getting better.


> Need to write a speech as a best man

> Write a eulogy for a deceased family member

Yeah, if you're okay with sounding like marketing copy written by a robot.

Remember last month when Stable Diffusion was supposed to put Disney out of business? Not gonna happen.


Amen.

It outrages me that someone would waste my time making me listen to a regurgitated, averaged-out, speech written without any emotional fire or depth at an occasion that is profoundly meaningful.

A speech as a best man should be personal, heartfelt, it should be funny and a little cringy maybe, and it should mean something.

Same with eulogies; this is a moment to celebrate someone's life, to state what that person meant to you, how they affected you, to share something about that person with everyone else who is mourning.

It's like those services that will buy gifts on your behalf for your "loved ones" or at least, the "acquaintences" that you seem to be obliged to provide gifts for. You've outsourced your taste, your chance to buy something quirky, meaningful, useful based on how well you know a person... which shows that you just don't know them at all.

Better to say nothing than say it with ChatGPT.


That's quite arrogant.

Not everyone has ever written something like it.

Why is it unfair or unpersonal if chatgpt helps you, guides you and gives you a good starting point?

Guess how many people would Google some example and starting with that, what's so wrong to let chatgpt generate a shitty first draft already tailored to your situation?


Hello greeny7373, nice to meet you.

> Why is it unfair or unpersonal if chatgpt helps you, guides you and gives you a good starting point?

As somewhat extreme examples: Rodin didn't swoop in and smooth off "The Kiss" after an apprentice chiselled the basic outline out of a lump of marble; pretty sure Salvador Dali started with blank canvas instead of getting a basic landscape from a $2 Art Shop and added a melting clock and a giraffe to it.

We all have language.

Take your age, subtract maybe 5 years, and that's how much experience you have expressing yourself in your language.

Say you're 20; by that metric you have 15 years experience communicating. Now, I don't play guitar and I'm not remotely musical, but I'm pretty sure if I did it every day for 15 years I could at least bash out somthing original, if possibly influenced by things I liked.

I'm not suggesting that people entertain us for half an hour with heartfelt, witty speech about their relationship with the groom; I'm not suggesting ten minute poetic ode to a life well-lived that leaves everyone simultaneous trying not to cry and trying not to laugh, and nodding and saying "yes, that's how they were".

It doesn't need to be long: just one memory or incident. It doesn't need to be Shakespeare: just heartfelt. It doesn't need to win the Academy for Best Actor. It doesn't need to be a 5 paragraph essay (though if you did use that model, there's nothing wrong with it).

It just needs to be you; it just needs to be yours.

(I already know what I'll say at my Father's funeral when the time comes: it'll be about two sentences, just something he said to me once. But I know that's all I need to say).

Maybe I am arrogant, but I want human feeling and human expression at human events. I don't want to be snoring through regurgitated pap.


I started to see a huge issue with knowledge: people don't know what they don't know.

If you have a magic Maschine your solution is no longer 'no clue how to do this' but it becomes 'i will ask my trusty expert machine'.

This machine now gives you the next step, the guidance you need without being jugemental OR much simpler: just looks to busy to you that you are not confident to disturb the other.


Some people are just bad at this kind of thing, and sometimes saying nothing is not an option if there is a social expectation that you say something. Which can create a lot of anxiety.


Yep. The best man's speech at my brother's wedding was a disaster zone in every possible way, it was a sad blemish on an otherwise great day. It'd have been far better if he'd have some AI help along the way.


I don't care how bad it is, I just want something from your heart.

And I'm sure you can still make something heartfelt with the help of a chatbot, but I'm afraid that people won't.


Many people won't with or without a chatbot. A chatbot might at least make it less cringe.


No one said Stable Diffusion would put Disney out of business but nice straw men


Really?

I'm too lazy to search for the Hacker News threads, but trust me, they did.

Though I agree, that was two months ago and we are so over that already.


> can take hundreds of pages of text and distill it down to an executive summary of any length

This is the kind of breathless claim that no doubt fuels the skeptics.

None of the context windows are large enough for "hundreds of pages" nor "executive summaries of any length".

I did believe that it's possible to make LLMs do that kind of task with significant engineering effort to make it do summaries iteratively somehow, essentially "compressing" parts of the document recursively. But it's not something that you can just give to ChatGPT and have it work.

So yes, the hype is real, in both ways: there's lots of potential to explore over the next years, but also a lot of the claims you read today are not sufficiently hedged, which makes them look outlandish.


There are plenty of tools that make it possible, such as langchain and the rest


Sure, quite likely - but only with significant engineering effort. I said as much in my comment.

The point is that it's not a turnkey solution today, but some of the comments on here make it sound like it is. That's a form of hype.


You describe it in a way that sounds like hype is negative.

Hype can also just mean a lot of media attention.

This doesn't indicate anything good or bad.


Unrelated to the general question of usefulness, but I'd be careful using it as a mathematics tutor. Maybe the algebra in this case is sufficiently basic as to be trivial for chatgpt, but I've found that quite often it gets confused by relatively simple mathematics in very baffling ways (One particular example I found amusing is it trying to use the pigeonhole principle with exactly N pigeons and N pigeonholes). There's also the problem that, being a complete pushover, it's hard (specially for a learner) to pinpoint whether the error is in the source or in the understanding of the reader. A teacher may, upon the request of a student to clarify what it seems like a mistake, say "No, this may seem wrong but it's correct for so and so reasons", but chatgpt usually goes "You're right, that was a mistake" and gets itself into messier and messier reasoning.


I don’t think you’re missing anything. I like the “calculator for words” analogy that was posted on HN a few days ago. It doesn’t seem like a revolutionary product, but it does seem like a fundamental innovation which will then unlock many more complex things in subtle ways. Calculators were arguably the predecessor to computers, which are kind of a big deal :)


> Calculators were arguably the predecessor to computers, which are kind of a big deal :)

"A computer is a machine that moves data around and only occasionally performs computation on it." ;-)

Computers aren't great, because they can compute numbers faster, they're great because we've managed to encode text, audio, video, geospatial, etc. information as numbers, which allowed us to perform complex text, audio, video, geospatial, etc. operations.



Most of the people asking the question "to do _what_, exactly" seem to have not much of a "knowledge worker" experience. Every non-tech person I showed GPT4 to was after a few tests immediately using it to speed up parts of their daily work.

Most impactful is how it destroys the blank page/getting started barrier, second how easy it is to substantially change/adapt/refocus produced work.

It is like having an incredibly efficient, patient and encyclopedic junior collaborator 24/7 at your disposal. It can't be trusted to fully automate without knowledgeable supervision, but it saves a boatload of time and effort.


> Most of the people asking the question "to do _what_, exactly" seem to have not much of a "knowledge worker" experience.

From personal experience I'd say the opposite: GPT lacks the specialist knowledge to produce useful writing or yield accurate answers in any of the markets I've worked in (I'll grant that less niche markets exist, and that GPT is pretty good at fixing the writing of people that lack English language writing skill) and it seems like people egging GPT as replacing most of those roles are all showcasing hypothetical "generate a website for an imaginary product with minimal brief" kind of situations which GPT excels at because there aren't any real world knowledge worker constraints imposed on their solution. That's definitely not to say it has no use, but lots of less technologically impressive accomplishments like data entry wizards and templates also have use without being considered transformative.


As is already a cliche now: GPT will not replace your job. Your colleague using GPT will.


As I already suggested: in my experience, certainty that GPT will be transformational in a field of knowledge work seems to actually be inversely related to experience of that field. A response whose certainty that GPT could be transformational to the stuff I worked on exceeded only by ignorance of what any of that stuff was is quite a good demonstration of that point...

(and really, there's nothing particularly special about any of the stuff I've worked on, it's just GPT doesn't have relevant knowledge or a path to acquiring it so doesn't generate remotely adequate responses, struggles even more with novel concepts and would be terrible at real time discussion even if suitable interfaces to it existed and were unobjectionable, and that's before we get started on the privacy implications)


> Most impactful is how it destroys the blank page/getting started barrier, second how easy it is to substantially change/adapt/refocus produced work.

This is the killer feature of GPT for me. I'm very, very good at optimizing and solving problems within specific domains, but terrible at picking a direction with no boundaries. (Pick a theme for a costume party and I'll have the most interesting costume. Throw a Halloween party and I'll show up in jeans and a t-shirt.)

I recently wanted to submit a conference talk, but wasn't sure where to start. I gave ChatGPT a list of possible topics and some general guidance about what I thought was interesting and asked it to suggest topics. I picked one from its list and asked for tweaks, then "discussed" with ChatGPT for a few more rounds until I had a very clear idea of what the talk would be.

I don't feel like that's cheating. I'm still going to create and give the talk myself. But if I had to come up with the topic and abstract on my own with only a blank sheet of paper to start with, I'd never have submitted it.


It's just good enough and approachable. But I agree, for what I do professionally it just doesn't even come close and I'm already quite efficient.

From me and around me:

- marketing asked me to show them the ropes of mid-journey yesterday, boss said "this will be the face of our new product" to one of my hasty "creations".

- mom learns english with chatGTP because she finished duolingo

- I wrote a PoC demo for a prompt engineering tool and gpt4 demo chat in a day. That would have taken me days without gpt4. (material design, storing in local storage, gobbling up the data specification, save everything on change not with buttons)

- First draft for some diagrams in mermaid js worked well too, or converting from flow diagram to a swimlane diagram.

- All kind of personal data cleaning: dirty list of emails > ready to paste in mail client, wall of text with broken new line characters > sub-headlines and paragraphs

- virtual assistants (e.g. company chat-/voicebots), needs a bit of tooling but gpt4 is totally ready for it (apart from latency and price)

- 30 minutes to a browser-add-on that marks tweets as "seen" so I can skip them if I scrolled past them before. (userscript to be precise)

- understanding tax regulations %)

What I'm waiting for:

- better knowledge ingestion so it can use my notes

- personalization over time

- good dev-ops integration (push and deploy for me too).

- maybe something to have better separation of concerns on code so the messiness matters less, not sure if possible :)

- remote control my screen

- running an LLM locally

- have it build its own plugins for any website or service


I‘ve been using ChatGPT to:

- write and explain me more optimised algorithms for certain cryptographic operations

- explain funny mathy bits of papers that I don’t understand

- plan me a few days of activities for a city holiday

So far it’s been great on all accounts!! I was able to get a faster turnaround time in understanding the papers than I would if I were probing a colleague


What gives you confidence its explanations are accurate?

> write and explain me more optimised algorithms for certain cryptographic operations

This domain in particular strikes me as a poor choice for this approach. "Don't roll your own crypto... but definitely don't let a language model roll it for you, either"


Well it gives me a direction to dig in - often papers use inscrutable notation or seemingly magical variables that I’m not sure where they come from. Will ChatGPT always be right? Probably not - but these are things I can validate better than no info at all!

Re: crypto algorithms, the quert in question was implementing exponentiation for arbitrary sized integers. My own implementation was taking until the heat death of the universe to finish for big integers and I didn’t want to just copypasta an impl from elsewhere.

ChatGPT‘s worked flawlessly and it was able to explain me certain tricks it used in depth (which I could independently verify from other sources).

Would I ship it to prod? Not without a security audit, but that ought to be the case regardless when rolling your own (or even someone else‘s) cryptography :)


This post is a clever joke, right?

Right...?


You never read something on a paper and had no clue how to even start to get it?

I talked about snippets of papers to work collueges. Chatgpt can do the same thing.

And why wouldn't it? It's only here to help me to find something I can use for further understanding/googling.


For a serious answer, my biggest reservation with using ChatGPT to explain concepts I know very little about is that ChatGPT often just makes up fake information, or very confidently explains things wrong.

Having to fact-check every thing it says feels a bit exhausting.

A bit like talking to a version of Albert Einstein with dementia or alzheimers, who can still say really intelligent things but mixes in subtle bullshit.


Well the alternative is scouring Google for answers and fact-checking them too - ChatGPT cuts out the scouring process


> I’m asking what people want the models to do every day, all the time

I know three regular people using ChatGPT, here's how they're using it:

1. Franchise consultant: uses it to research opportunities, has it write business letters to franchisees he wishes to contact. Saves him time and is a better writer than he is.

2. Immigration lawyer: uses it to summarize info and write emails to clients. Saves her a ton of time.

3. School teacher: uses it to write report card and assignment feedback. It doesn't save any time at all, but the output is more elegant than if he wrote it manually.


You are overlooking an important use case: SPAM.

Spam of all kinds and at automated-industry scale. Imagine blog-spam written in a variety of styles so you can no longer easily identify it as such. Imagine chum-boxes that no longer repeat themselves and are harder to identify as such. Imagine ads masquerading as content, as it already exists, but at scale.

And every time you slip up and click on one, it will learn a little more about you and create chum content better tailored to you.

Generative AI will facilitate a flood of algorithmic spam.

This kind of spam already exists, it just isn't scaled because it still takes time and resources to create and the people creating it are not the brightest, so for now it is easy to identify.


Natural language is now a fully functional user interface out of the box. This is bigger than the mouse.


Fully functional in what way? As far as I can tell, ChatGPT is a box that I put sentences into, and I get grammatically correct sentences that contains topics or words loosely statistically correlated to what were in my sentences, that may or may not be correct and often are not. The box has little to no memory. I honestly don't see what's so useful about this box.


The results are approximately as good as asking an intern, but the response is approximately instantaneous, where an intern takes a week to do anything. And similar to the intern, you can get better results with a bit of guidance and iteration.

In short, humans kinda suck. LLMs also kinda suck, but faster than humans.


> humans kinda suck

Agreed, but that's part of my critique. These systems are written by humans and are trained with barely curated data generated by humans. There is the concept of emergence, but I'm not sure how emergence suddenly fixes a terrible foundation full of biases and errors.


> and I get grammatically correct sentences that contains topics or words loosely statistically correlated to what were in my sentences, that may or may not be correct and often are not.

It's correct enough to pass the bar exam, medical exams, it scores 90-93 percentile on the SAT. This is way more complex and efficient than what you make it to be imo.


> It's correct enough to pass the bar exam, medical exams, it scores 90-93 percentile on the SAT.

So is Google Search, and we had that for a long time now. Is a slightly different and more verbose UI really a game changer?

(Let's ignore the fact that Google Search is broken from all the SEO spam and monetization. Especially when we have no evidence that ChatGPT is any more resitant to this than Google was.)


Google Search scores 93% on the SAT? On questions that have no solution online?


When hooked up to plugins or tools it does feel like a fully functional interface. With access to a browser, it can do basically whatever I want it to.

I hooked GPT-4 up to a shell and asked it to use the GPT-3.5-turbo completions API (not in the dataset), and it successfully did it through trial and error with curl and error messages. This example is of course not something you would actually do regularly, but rather shows that you don't need a lot of context for it to do useful things right now. With a complete explanation of the OpenAI endpoints, it would most likely make the request perfectly on the first try.


Yes, exactly this is what I meant. You can really leverage the API in other systems. CharGPT on its own is cool, but my mouse comment was about NL being a bolt on UX tool. Which is crazy!


You can tell it what you need to do in terms of data processing, and ask it to write Python code that does that. E.g. ever had to write a convoluted ImageMagick command to do something complicated? GPT will write it for you.

Think of it as a natural language interface to anything that has an API.


You can teach it to act as a support personal.

'only answer the following questions and use these API endpoints doing the task'.

Now you have a multi lingual human Isabel Interface.

Have you seen the Wolfram alpha examples?


I think the problem is that people working in the field have a much different definition of what "fully-functional" means than users who believe that they're communicating with an intelligent, all-knowing or otherwise infallible being.


For my use case I use it for input processing, it’s just amazing that way. Hallucination is a huge problem, but you can avoid it.


I believe it. For text->text workflows it's a game changer. For anything touching a physical systems, even the highest-level path planning for robotics, there's a long way to go.


It’s all about being very deliberate with your use. People are getting ahead of themselves right now for sure. It’s still a monumental leap in UX!


In the long run, it will 'secure' knowledge and data. Nobody will know what's in GPT's databases.

Currently you can command kagi/google/http websites to return information. You can infer what should be in Google's search engine and track when information is deleted.

GPT is not commanded, it predicts with inaccuracy. So anybody who wants to black-hole information behind the scenes and never reveal clues to that fact, can do so.

All failed predictions are covered by LLM's design, you cannot infer without serious long term study that something has been removed deliberately. You cannot infer that a valid data entry exists and you failed to retreive it, because unverifiable bs is the default failure state from LLM's.

High level tech people will invest in this, regardless of what the public values in it. Just like Elon's SpaceX and Tesla got lifted out of pitfalls by gov and VC, so too will the AI guys.

Let me put it this way. Hoarde and backup every scrap of online information you care about. Hypothetically an LLM fueled replacement for all of 'the free and open web' websites, could limit information availibity.

A metaphorical example would be leaving out Tianamen Square. Which is fine when you can just Google it, but with the old freedom of information gone, an LLM has the ability to just bs you and you'd never have a reason to infer it existed in the first place.

It's a Super-Injunction by default, a perfect repository for spies to dump data, a librarian who will answer any question but only answer with the truth, if he likes you.

No more Snowden and Assange leaks, there's no way to chase up a deleted video with a search engine.

Anyway you get the idea. In the long run, the structuralists are licking their wet lips at the thought of re-establishing a heirarchy of information access. (Probably, i don't know).


> I’ve seen examples of “write an email to my boss”. It would take me longer to explain to ChatGPT what I want than to write these myself.

You’re a better writer than me also English isn’t my first language.

Besides that I used chatgpt to write a eulogy for my father. I know roughly what I wanted to say but I couldn’t find a good way to say it. Chatgpt helped me there even in my own language (Croatian) and there is just no way I could’ve made it as poetic as it did.


For me it’s a perfect replacement for Stack Overflow. Except every solution is tailored to my exact code and situation. I’ve even gotten it to walk me through things like installing WSL 2 without using the Microsoft Store after I nuked all Appx packages.

Maybe my favorite was pasting in pages of documentation describing all of the error codes for a library (the docs are the only source of truth) and getting it to output a very good typescript enum.


You don't need a full-blown LLM to find answers to questions on StackOverflow. You can do that today with any search engine.

> "I’ve even gotten it to walk me through things like installing WSL 2 without using the Microsoft Store after I nuked all Appx packages."

That's a search engine query you can issue already today, without involving any LLMs. It will cost a fraction of the cost of running this with an LLM, and it'll actually bring you to the "source" of the information (a thread on StackOverflow with the full context - including "wrong answers" which are just as useful) - unlike an LLM.


The problem with SO is asking questions though. It’s a horrible site for that purpose. I rather get a slightly wrong answer from ChatGPT than the insane experience of trying to formulate my question just right so it won’t get closed or deleted on SO.


I remember cross referencing with Google to see where ChatGPT was pulling from and it wasn’t immediately obvious. So yes I probably could have found a guide through a search engine with more digging. But clearly the LLM is the superior product.


[flagged]


Sure, you can keep throwing those hyped VC/MBA-style retorts all around, or you can acknowledge the simple reality that when it comes to the task of finding answers to questions on SO, traditional search technology outperforms LLMs on every parameter:

1. Cost: it's much easier to build and serve a traditional search index for questions/answers at scale

2. Speed

3. Quality: with search engines you don't just get one answer (you get a spectrum of answers). You have access to the "source" of those answers, which often include additional data that can help you choose the most appropriate solution to a problem.

In this specific case, the "horse" is the LLM - and the "car" is the traditional search engine with modern indexing technologies. But please, don't let simple textbook facts about technology take you off your trip.


I'm quite good in understanding the difference.

But LLM can already do things Google can't.

LLM is not just a search engine.

It's an explainer, rubber duck etc.

Instead of 'google define x' I just say ' explain x' in my dialog with the machine.

And it will only become better and better.


Probably because much of the time it's just a (hopefully correctly) curated subset of stack overflow ... tailored to your current focus.


Fine with me!


Please write a blog post about this! That is an amazing story.



Which one?


I have a friend that has been using ChatGPT to help him write technical books which collect existing niche information and present it in a way that's more accessible and organized, in Spanish. You'd be amazed by how many small niches there are without robust technical books and well organized resources. And while the information is all "out there", it's often not really easily accessible to beginners.


Text redaction - if any part of job is about writing, you can now just brain dump everything, structure it more or less, throw into chatgpt, and make it produce a clear and readable article. In my case, the output is better than anything I would ever write.

Ditto all the redacting work in newspapers, intranet etc. The whole field of proofreaders was virtually extinguished overnight.

Marketing agencies - and I spoke to a few - increased their workers' producitivity 2-4 times (sic!), virtually overnight. Anything from writing briefs to writing copy.

Programming - most of my programming work is deep algorithms, so not much help here, but for writing boiler plate code with new APIs, or writing in a language that I'm a bit rusty in, chatgpt is better than anything else.

Customer service helplines / chatbots (and the same for intranet) - we don't see it just yet, because it takes a bit more time to build a good system, but there are probably thousands projects right now worldwide building those for any niche concievable.

Business intelligence - we used, with success, ChatGPT in our deep tech seedfund, for helping out with initial project ddil.

And, essentially, rubber ducking, but for every single field out there. I just discussed with a psychiatrist how he can use even boilerplate GPT-4 as an additional consultant. You need to be aware of limitations of course, but it is already immensely useful in it's current form - and dedicated solutions for medicine are coming very soon.

That's the short-term perspective and low hanging fruits. On top of that, you have thousands projects now, that are figuring out how to apply LLMs to specific niches. It was difficult before, because you had to train your own models, and now you can just fine tune the existing ones, do embeddings, or just plain prompt engineering.

Oh, and also synergy with different AI modalities - we've had a massive growth in voice recognition and generation, visual recognition, and so on. And LLMs are a glue that adds a layer of understanding underneath.


I hope you discussed data sharing confidential patient health information with third parties, openai etc.

Nevertheless offline models should alleviate some of this.


Yes, I did, of course :)


"In fact, no one in my extended sphere of friends and family has asked me anything about chatGPT, midjourney, or any of these other models. The only people I hear about these models from are other tech people."

I have the opposite experience. Everyone I meet outside tech has been exploring ChatGPT, or at least has heard of it and is extremely curious. And in non-tech student circles, a (non STEM) TA I know said students who until a month ago were bad to mediocre (8-12/20 scoring), are suddenly all turning in top 16-18/20 assignments this month. You can argue about what to do given this change, but you can not deny the impact.


I have the same experience. And ChatGPT was the turning point. A few month ago, to my surprise i wasn't even able to excite my non-technical environment about DALL-E 2, not to mention other ML models, that would excite technical people.


I was sitting in a Starbucks yesterday sitting opposite two girls doing some language study and one of them was joking that the other should ask ChatGPT about some grammar point. This idea that only tech people know about it seems way off.


We are going to have almost on the fly translation of video and audio. We are going to understand and write code faster. Knowledge management will change so instead of being partly a tool which searches the web we will get human readable responses. Language model is about to improve language related things for us, how killer it could be?


Code things i've used it on: - I've thrown in jsx react components and asked to transform it to tsx - I've thrown in tsx components and asked it to write some tests - I've asked it to make an offline queue's so user requests are stored and saved once online again - I've asked to rewrite tests from enzyme to Jest - I've asked it to add a full screen function to react-native-video, which is annoying.

Often Im doing something else in the meantime, like coding, in a meeting or having a beer.

Finally we can be drunk and code at the same time ;).

Joking aside, it's been a huge productivity boost, and if you ask things properly will write hugely detailed and correct code.

I've also used to understand other languages/coding I was less familier with, for instance c and sql procedures.

Above only works with v4, 3.5 is to inaccurate. But is indeed slow, small things I can do faster.

With writing articles i've been disappointed so far, even corrections or styles I ask to adjust it puts in back in a few questions later.

However writing children stories for helping my kid learn how to read it's really good, in every language. "Write a story for kids of age ... about ... use the following words". Came up with nearly perfect stories my kid loved.


I use ChatGPT at least 10 times a day, as a software developer, and I agree with all your assessments about the use-cases you mention. Here are my ten most recent prompts:

* Asking about how back-propagation works with multiple output nodes

* Examples of successful ICOs

* A deep dive into what it means to calculate the gradient of a function, with me asking lots of clarifying questions

* A deep dive into how electricity pricing works in the UK, digging into the market clearing price

* Looking for a generic "unwrap one layer of type" utility type in TypeScript

* "In JavaScript, I want to format a date as YYYY-MM-DD_HH-MM-SS"

* Seeing if there's a more concise way to get Zod to define an item that's required, but the value can be defined, other than using a union

* Ideas for naming two user fields, one that's a changed user, one that's the user making the changes

* Digging into the implications that Camilla is called the Queen Consort, not just the Queen

* "Excel I want to show the weekday as a single letter"

So, basically, it's my go-to instead of Google + StackOverflow


I’d be interested to see what it said about differentiating a function (and I’m assuming you mean in the context of programming). I consider that one of the dirtier secrets of gradient-based ML.


Also I've found it very useful for identifying books, TV shows, films, games, etc that I remember some details of (eg, was released around 1994, was about web design) but not all the details of


You, good sir, have a disturbing lack of imagination.

What I can ask my computer has just leaped from templated strings to raw human conversation.

I have a hard time imagining something not getting impacted.

First of all iPhone autocomplete.. (but I guess any decade old RNN is an improvement there)


It's just a glorified google search. You'd save more time googling what you need and picking 2-3 results to read briefly.

People have no clue human intelligence has very little to do with "statistical analysis of old data".


It's actually quite a lot better than a Google search. Yesterday I gave it the output of SQL ddl and told it to write me a script in Go that reads that from a CSV, then I asked it to generate a sample CSV to verify the script.

How is that different from googling? Lots of articles won't bother to include import and the structs they use definitely won't match up to your use-case because their dataset if different. What if my use-case is a bit strange and I need to embed the file instead of reading from the file system, for example? I can ask Chat GPT and it will update the example program, using my exact file name and the variable names for the problem I've described and the program runs as written!

I don't write much Go (so I don't actually care to commit the hello_world.csv ritual to memory), but I know enough about it to verify that the program doesn't have any glaring issues and make my own tweaks as necessary. Saved so much time for me in this scenario.


It can't replace search, actually. It doesn't know everything a decent search engine can find, and might not even remember the specific php command line parameter you're looking for.

What it can do is reason some pretty tricky interdisciplinary answers. It's about as useful as having an intern that's finished literally every white collar college out there but has no idea what they're doing. It takes some prompting to get useful work out of it but it is possible and it is very effective when you get it right.

The $20 sub has saved me quite a bit of time, the main challenge is remembering to set it to GPT4 for every damn conversation, as the previous models are trash.


I use 3.5 because 4 is too slow and hangs half the time.


Google search is a trillion dollar product.


I have read a lot of books (sociology, psychology, philosophy, history, religion, etc), and often I'll have an insight that stands in apparent contradiction with some concept I learned in one book but in alignment with a concept I've learned from another.

I use ChatGPT to understand these concepts, their background, and to bridge the gap between them.

Getting to the bottom of what Jean Baudrillard really was saying in "Simulacra and Simulation" and applying that to what Larry Wall meant when he said that Perl is the first postmodern language is really something!


similarly what’s the point of stackoverflow or wiki when you can just get the answer you want from a book in the library.

What I value about these lls is that they are essentially a condensed version of the internet (despite being stupidily large for normal hardware currently)

Usually if I’m building a recommendation or search algorithm I have to use the data from the company I’m working with. This makes it possible for me to encode the entire internet into a model that might be running on a product that only has 100 users.


LLMs will make average and below average writers / marketers / PR to become extremely good.

You no longer can mock "Please do the needful" from $1 / hr employees from India. They will be communicating at the same level as an average American and the smart ones can completely take over large fields.

A single person, can wear multiple hats and not get blocked.

This can apply at global scale of at least 4 Billion.

Just because you lack imagination doesn't mean it isn't good


> I’ve been shown some neat pictures people made that they thought were cool. I don’t know that I need this every day.

At least they’re less obsessed with apes than last summer…


To be honest, I can't tell if you are genuinely wondering or just (virtue) signaling that you're not using it, but here's a sample of my own usage: https://news.ycombinator.com/item?id=35299121


Not virtue signaling, but I seem to be asking a different question than people are answering. I’m asking what the product opportunity is and people keep telling me examples of tasks that they use it for.

In many cases the examples are one-off and the only product opportunity is the generative model interface itself. Looking broadly over the replies what I’m seeing is “there’s a thing that used to fail the cost/benefit test, but now the cost is so low that I can automate these things”. So part of my problem is (1) the small benefits of these tasks mean the value proposition comes from volume—that probably comes from the generality of the task engine, and (2) there may be some niche product opportunities on top of the model platform, but the primary big winner here is the platform itself. (That’s not necessarily a new insight, but it seems especially true here.)

The terrifying part is how often I hear people in this thread and elsewhere mentioning tasks that are not fault tolerant to the failure modes of these models. (For example, I had a coworker tell me their relative is a doctor using ChatGPT to diagnose patients.) People keep focusing on the risks of AGI killing us all with paperclips, but I’m much more worried about getting run over by some idiot asking ChatGPT to drive their car.


> Not virtue signaling, but I seem to be asking a different question than people are answering. I’m asking what the product opportunity is and people keep telling me examples of tasks that they use it for.

Sorry, I misunderstood. "I am not using AI" has become a sort of badge of honor in certain communities so I was wondering if that's what it was.

> (2) there may be some niche product opportunities on top of the model platform, but the primary big winner here is the platform itself. (That’s not necessarily a new insight, but it seems especially true here.)

I agree with that conclusion. I think the chat interface is the killer product. I treat ChatGPT as an assistant/intern that is really good at some tasks but that can also sometimes make dumb mistakes. It has also replaced a lot of queries I would have previously done on Google or questions I might have asked somewhere (e.g. in a forum, Reddit, Discord, etc.).

Many startups build domain specific UIs on top of it using the API, but whether that will become a sustainable business model remains to be seen[0]. I am reminded of the many "vertical" search engines that were once trying to compete with Google.

[0] Saying this as someone who did something like that: https://eli5.gg


For me, it gives an inkling of a web that existed before, with less friction. Instead of going to a webpage with verbose ad-ridden fluff, I get a more or less frictionless answer, very specific to my question.

Funny enough I don’t use it a lot for programming, maybe just to jog my memory on a topic.


The thing has enabled me to upgrade a major code project into an exciting direction in a short amount of time. Paradoxically I find it does better with more difficult tasks because the prompts tend to have more detail. the thing can generate semantic cypher queries for combinations of combinations of combinations of elements if you want, it’s ridiculous at coding already and only gonna get more insanely good at it in the future. Yeah, sometimes it forgets context and you have to guide it a little, it still can rewrite a whole module in a new style with lots of specific changes under 2 minutes, good luck coming close to typing that fast even if you did know exactly what to write, coding feels pretty revolutionized overnight right now for me tbh


I don't have the same experience as you at all. Most of my non-tech friends or family members asked me about chatGPT. You don't find it useful to write emails maybe because you're a native speaker, but it is extremely useful for non-fluent english speakers.


I defined an interface and asked ChatGPT to implement it, in Go, as method receivers on a struct named “repo” against sqlite. The code compiled and I wrapped it in unit tests to ensure it worked. Overall saved me an hour of boilerplate, would recommend.


What is the killer app of the internet, or the smartphone? It's a silly question. It's many little things, some foreseen and some not.

One thing you're missing is that we now have a pretty good solution to any NLP pipeline that in the past you'd have to spend months to get right. You can probably still get better results by supervised training on specific tasks but it's good enough. NLP (as we knew it) is dead. This will take some time to show in the applications we use, as people figure out how to use and integrate it, and costs need to come down, but it will make it trivial to add smart functionality for things you previously needed an in-house ML team for.


Sorry, one other thing, as an actual AI researcher:

These generative models, whether NLP or vision, are cool but are really examples from a very narrow field. Most ML researchers and practitioners are working in completely different areas and would not obviously benefit from the new generative models (which are themselves prodictized extensions of existing tech trained on more data) so nothing's going to change day-to-day. If you were working on some alternative general purpose generative model, maybe you got scooped. Otherwise it's business as usual.

The "revolution" is happening for tech savvy non-ML people, "tech-bros" colloquially.


I'll ask you another question: What are people searching for on google right now that can't be answered by ChatGPT? I would say 95% of searches can be a simple question & answer of ChatGPT. I've tried it, it works.


Note: When I write "Midjourney" below, I mean any AI/ML-based art generation tools.

Art generation is going to be groundbreaking for the advertising industry. Basically, you can hire summer interns from arts academies or liberal arts schools to type cool stuff into Midjourney to generate amazing art for your ads. You don't need to pay for real artists, who, sadly, their original artwork was used to train the model.

The same can be said for low-end graphics arts: Stuff like: Make me a greeting card for this big event. Today, you need to pay a graphic artists to whip something up in 2-72 hours. That will be replace by someone with modest English skills working in a call center somewhere in India or Philippines. They might chat with you for five mins (voice or text chat) to brainstorm ideas. Then, they will "drive" Midjourney and put together a nicely-themed party invite.

On the more advanced side, I do think artists will browse the "best of" portfolios on Midjourney (and others) to get new ideas. They might also use Midjourney to get a head start.

The next logical step beyond Midjourney is to generate the same image, but as a 3D model. I think people (myself included!) really (really, really!) underestimate the cost of creating 3D models for films, adverts, and games. If Midjourney could give you a starting model, you might save hours (or days) of work.

Next-Next: Midjourney can provide basic animations of the same 3D model. Again, you can download in a wide variety of formats, so that you can import and tweak as necessary. Think: Pay to play. Rendered as GIF is cheap as hell, but download advanced CAD format with 50K points in the model might cost 100s of USD. (Remember: You are paying for a SaaS engine, not an expensive, talented artist.) Imagine: "Hey Midjourney, I need a 60 second animation of cute animals, like Animal Crossing, sitting at a table enjoying our new brand of tea called 'It's Great Tea Meet You!'." (Use ChatGPT to generate the first draft of the script.) Writing that just made me think: Ok, now you can add voiceover.

The possibilities for the commercialisation of generating still and moving picture are nearly endless, and many will be useful for the advertising, film, and gaming industry.


I think the killer app will involve adding more layers of AI to make something that is closer to an agent than a simple prompt completion engine. Maybe a more self-contained system that has an internal cycle of asessment, generating prompts, executing output, and readjusting prompts. These agents could conceivably work on long term, complex projects with minimal supervision or interaction. If a general purpose system can get a foothold and be even 10% as good as a person, it could be enough to fundamentally reshape the workforce.


One thing I find it really great for is language learning. It can create examples, explain grammar, contrast to languages you already know etc. Also I'm not a sys admin but occasionally have to do sys admin things and I ask for the shell commands I need in English, it is far far faster than using stackoverflow etc. (i.e. pulling logs from multiple servers and extracting bits) Also I don't want if I want to use a library I ask for what how I want to use it and get a small example entirely tailored to me.


My girlfriend and I use it to turn a short bullshit paragraph of corporate lingo into a bigger one. It's great for that, but that's it.

GPT will only ever be a good writing support tool.


In the content space. There's whole publishing industries that are at risk from GPT style models.

Would reddit, or HN still be interesting if there were bots that talked in the popular tone of the subreddit or thread that dominated the conversation?

AI's effect will be like pollution to our culture. It'll speed up and help alot of things and be very helpful generally. But It will create problems in 'human' spaces.


> I’ve been shown some neat pictures people made that they thought were cool. I don’t know that I need this every day.

3d and motion and this becomes the holodeck, you describe a scene and it creates it direct to your vr goggles.

Generative AI that can take in a scene and alter it rather than fully create it becomes real augmented reality from sci-fi: not just a HUD or greenscreened in elements, major transformations.


There is no killer app because chat gpt is largely useless. It's results might be right, half right, half wrong, or totally wrong. All it's good for is spouting bullshit. It's like talking to a really outgoing, really confident used car salesman.


I’ve got a codebase written in the extreme programming paradigm that has become a linchpin for operating our enterprise.

I can fire up an AI to build out my missing unit tests in 10 minutes what would take a developer 3 weeks to accomplish.

Scale that for 155 different projects … that’s over a year of development time in about 24 hours of compute


Imagine a code/progtamming AI that I can ask to code review my code and find my mistakes, write unit tests for my new functions.

I am tempted to write a small program to use to fix my Internet comments and fix my bad English exprimattion , I done some tests and I see that ChatGPT find the places in my comments that can be improved.


Instead of looking at use of an LLM in isolation, consider the value it could have in the future, combined with API's, other models, speech interface, etc.

Sure, it can help you write an email now, but the real magic could be when these things come together in a symphony of intelligence.


Yeah it is tiring, currently.

But soon people will just enter one goal, "make me money" and the program will go on a loop, pausing only when making an important decision to get the approval of the owner

Also, the input buffer of LLMs is increasing. Soon we will be able to begin with "write me a full 3D game"


> In fact, no one in my extended sphere of friends and family has asked me anything about chatGPT

A couple of weeks ago they wrote an episode of South Park with ChatGPT (and it was about ChatGPT). It's definitely gone mainstream amongst students who are using it to do homework.


I use it for APIs. Specifically, I use it to help think about how I should architect my APIs, as well as how I should implement them. I can ask ChatGPT to do both things for me. Even if it's somewhat wrong, it still gives me a good enough direction to go on.


Almost all office work is just reading words in boxes and looking at pictures and then writing words in other boxes.

There's insane potential to automate it. Think b2b, not b2c.


“Almost all” makes me question how much experience you have doing those jobs. They exist but work which is truly that rote has been getting automated for half a century so what I’ve found is that what you see now tends to deceptively appear simple while hiding judgement, uncertainty, and inference.

That’s especially true for things which involve liability. If Google builds a system to recommend YouTube videos and ads, it’s a win as long their error rate is below a certain level. If it’s your insurance company rejecting claims, however, people can die that way and the lawsuits for breaking contracts or legal standards can far exceed the believed savings.


Killer app is that it's glue natural language code. It's not realised yet but it's gonna be BIG. It's simply gonna be copilot for everything.


Have you heard of "editing"? Or "translating"? Those are entire industries that can now be almost entirely automated with no loss of quality.


It replaces googling. All the hard work of finding relevant info and piecing it together is done for you. For programmers it's just a faster stackoverflow.


I have a group of nontechnical friends who use Midjourney constantly for D&D and similar games (both brainstorming and visuals for players)


Ask college kids. They’re all using it for everything.


For a non native English speaker it's amazing. It can take any very broken blob of text and turn it into perfect English.


I use it to plan meals and make grocery lists each week. It's very good at that task, and it saves me a lot of monotony.


> I’m asking what people want the models to do every day, all the time.

Rule 34. When in doubt, always defer to rule 34.


Me? It takes the sql for dozens of DBT models and spits out the schema yml so I don't have to.


Parent article said the pressure was "be the first or be the best."

The proper answer is "to be correct."

Thousands of Africans sat at their desks, trying to achieve correctness.

This will become a (correct) commodity, but not today.

(Unless you are really good at differential equations.)


Scrolling back through some discussions with chatgpt and a few things with bing (later ones more likely to be gpt4)

* Explaining what a do/while(false) loop was for as I'd not seen that construct before

* Discussing what DIDs were and how web DIDs worked as the RFC was very detailed in a lot of areas I just didn't care about. A discussion with a pretty knowledgeable person in the area was what I needed, and what I got. It explained what the well-known part was, explained my confusion around resolving to a document and resolving to a resource (where I was mentally stuck).

* Creating a learning plan, diagrammed by mermaid, for progressing in bouldering. Each major step broken down into sub-parts to practice

* Finding https://en.wikipedia.org/wiki/Bloom%27s_2_sigma_problem given bad explanations that didn't lead me to the right place in google

* Finding https://blockprotocol.org/ given a shaky memory of "some composable ui framework by someone famous" iterating a little with bing. I had failed to find it before manually.

* Explained and created a table of bouldering gradings as the place I go uses a different one to the videos I see

* Discussed project ideas to do with my son, gave me great ideas around electronics that I think are a good fit as well as a few other things. The most useful part here was being able to say "that's too generic, I need proper projects" and "that's too simple for him" and have it update. It then also created some good explanations at different levels about how radios work,

* General discussions about long term impacts of LLMs, potential use cases

* Career advice

* Generating art that we'll be getting commissioned for the house

* NER without any coding

* Generating ember templates, CSS and example filling data for a custom framework given a problem statement (what I'm actually building right now)

* (edit) I just took requests from my kids and made them some colouring in pages with robots, firefighting robots, lego ninjas, owls, frogs and crabs.

> In fact, no one in my extended sphere of friends and family has asked me anything about chatGPT, midjourney, or any of these other models. The only people I hear about these models from are other tech people.

I've had family members tell me they've used it to create reports, and used it to create marketing copy, a website and lecture slides for others.


Needed to safely move a bulk carrier on short notice recently:

Q: Describe turning 180 degrees and continuing down stream with the current in a heavily loaded bulk carrier from a stationary position tied to a dock with five knots of downstream current using two 50 ton Ballard pull tractor tugs. Be very detailed and specific in describing the actions required to complete this action safely.

  A: To safely turn a heavily loaded bulk carrier 180 degrees and continue downstream with the current from a stationary position tied to a dock with five knots of downstream current, using two 50-ton bollard pull tractor tugs, you must follow a series of steps. These steps are designed to ensure a smooth and controlled maneuver while minimizing the risk of collision, grounding, or damage to the vessel, dock, and tugs.

  1. Preparation: Prior to beginning the maneuver, ensure that all necessary communication equipment is functional and tested. This includes VHF radios, signal flags, and/or lights. Establish communication protocols and channels with the tugs and any other relevant parties, such as pilots and harbor control. Additionally, check the weather forecast, tidal information, and any local regulations or restrictions that may apply.
 
  2. Positioning of tugs: Position one tug at the bow and the other at the stern of the bulk carrier. The tug at the bow will primarily be responsible for pulling and steering the vessel, while the tug at the stern will act as a brake and assist in maneuvering the vessel as needed. Securely attach the towlines to the designated towing points on the vessel.

  3. Cast off lines: Once the tugs are in position and ready, instruct the crew to release all mooring lines from the dock. As the lines are released, the tugs will need to maintain tension on their towlines to keep the vessel in position and prevent it from drifting downstream with the current.

  4. Initiate the turn: The tug at the bow will start the turning process by pulling the vessel's bow away from the dock and into the current. At the same time, the tug at the stern will push the vessel's stern towards the dock to create a pivoting motion. Both tugs must continuously adjust their power and positioning to maintain control of the vessel as it turns.

  5. Manage the current: During the turn, both tugs must work together to counteract the effects of the five-knot downstream current. The tug at the bow will need to pull harder to overcome the current, while the tug at the stern will need to push and also act as a brake to prevent the vessel from gaining too much speed.

  6. Monitor the turn: Throughout the 180-degree turn, the vessel's crew and tugs must maintain constant communication to ensure a smooth and controlled maneuver. Monitor the vessel's position, speed, and proximity to the dock, other vessels, and any navigational hazards. Adjust the tugs' power and positioning as needed to maintain control and avoid collisions or grounding.

  7. Complete the turn: As the vessel approaches the completion of the 180-degree turn, the tugs will need to adjust their power and positioning to slow down the vessel and align it with the downstream current. The tug at the bow will ease its pulling force, while the tug at the stern will continue pushing until the vessel is properly aligned.

  8. Proceed downstream: Once the vessel is aligned with the current and moving downstream, both tugs will gradually reduce their power to allow the vessel to continue under its own momentum. The tugs will remain close by to provide assistance as needed until the vessel is safely clear of the harbor and any navigational hazards.

  9. Release tugs: When the vessel is at a safe distance from the dock and navigational hazards, and under the control of its own propulsion, the towlines can be released, and the tugs can return to their base.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: