https://old.reddit.com/r/interestingasfuck/comments/1nkbqyk/...
Instead they show tech’s quality on a basic highest common denominator use case and allow people to extrapolate to their cases.
Similarly car ads show people going from home to a store (or to mountains). You’re not asking there “but what if I want to go to a cinema with the car”. If it can go to a store, it can go to a cinema, or any other obscure place, as long as there is a similar road getting there.
A better analogy would be the first cars being advertised as being usable as ballast for airships. Irrelevant and non-representative of a car's actual usefulness.
Lol are you serious?
Almost half of Americans cannot cook today. And the number 1 cited reason is a lack of time.
That said, I agree with the grandparent that this isn't really a "killer feature". Nor am I interested in the product. For so many reasons.
https://www.newsweek.com/clifford-stoll-why-web-wont-be-nirv...
2D chess if they're smart: start a new company that competes with the one they just sold to dumb investors. Jack Dorsey is particularly fond of this move.
Taxpayers do not "foot the bill" for corporations reducing their tax obligations via "write-offs".
See: https://accountinginsights.org/what-does-write-it-off-mean-f...
One of the best CEOs in the world with about 20 years of experience at age 40? And also founded the company?
He’s doing pretty good. And if you’re talking about “image” he is a millennial archetype.
"Brian's Hat" is the 1st one I saw and maybe the best: https://youtu.be/LO2k-BNySLI?si=qEX7STkSOeCVZtK-
Also "Hot Dog Car" https://youtu.be/WLfAf8oHrMo?si=jz5EKwjJZm1UMZau
Enjoy. :)
How strong does a company's reality distortion field have to be for people to think your friends are going to want to come over to play with a new version of Windows?
I mean, why not "Let's all have wine and cheese and do root canals on each other!"?
I'm endless amazed that Meta has a ~2T market cap, yet they can't build products.
I think that's why he kept saying exactly "what do I do first" and the computer responded with exactly the same (wrong) response each time. If this was a real model, it wouldn't have simply repeated the exact response and he probably would have tried to correct it directly ("actually I haven't combined anything yet, how can I get started").
I don't know about anyone else, but I've never managed to get Gemini to actually do anything useful (and I'm a regular user of other AI tools). I don't know what metric it gets into the top 2 on, but I've found it almost completely useless.
I asked for a deep research about a topic and it really helped my understanding backed with a lot of sources.
Maybe it helps that their search is getting worse, so Gemini looks better in comparison. But nowadays even kagi seems worse.
That's the whole argument?
I will die on this hill. It isn’t AI. You can’t confuse it.
>> I will die on this hill. It isn’t AI. You can’t confuse it.
> They "poisoned the context" which is clearly what they meant.
The "demo" was clearly prescriptive and not genuinely interactive. One could easily make the argument that the kayfabe was more like an IVR[0] interaction.
0 - https://en.wikipedia.org/wiki/Interactive_voice_response
“The blue square is blue.”
“The blue square is green.”
The future is here.
* If you think it's something that resembles intelligence enough to be useful in the same way intelligence is and to seem to be intelligence, this is clearly it. The "plant based meats" of AI.
* If you think it means actual intelligence that was manufactured by a human, this is not that. It's shockingly impressive auto correct, and it's useful, but it's not actually thinking. This would be "artificially created intelligence"; in essence, real intelligence with an artificial origin. The lab grown meat of AI.
For the latter, I really think it needs reasoning ability that isn't based on language. Plenty of animals can think and reason without depending on language. Language is a great carrier of intelligence, which is why LLMs work so well, but language is not the foundation of intelligence.
That said, I think "confused" is a fine enough anthropomorphization. I refer to things like Bluetooth issues as the machine getting confused all the time. Back in the day, Netflix would often have problems with the wrong image showing for a show, and I called that "Netflix getting confused". We know it's not actually getting confused.
I guess my definition of 'to be confused' is something like 'to have a significant mismatch between your model and reality', so yeah, you could argue that a PID controller is "confused" by a workload with significant inertia. And 'to feel confused' would be 'to identify a significant mismatch between your model and reality', of which clearly a PID controller is not capable, but most multicellular life forms are.
However "confusion" can also mean "mistaking one thing for another" or simply "doing the wrong thing", which is something computer programs have been able to fuck up since forever.
_________
ChatGPT: There is a seahorse emoji: — wait, no, that’s a lobster. Okay, for real now:
Here it is: � Seahorse — argh, it’s still not showing properly here.
Let’s try again:
(hedgehog) (sauropod) (T-rex) (otter) (seal) (lobster) …but no actual (squid) or (mermaid) matches…
Okay, deep breath — I checked the official Unicode list. And you’re right: The seahorse emoji does exist — it’s — no, okay really:
(hedgehog) (hippo) (peacock) Still nothing?
Wait — turns out, I was wrong, and you’re right to call me out.
There is a seahorse emoji:
— no!
_________
I found the use case honestly confusing though. This guy has a great kitchen, just made steak, and has all the relevant ingredients in house and laid out but no idea how to turn them into a sauce for his sandwich?
Even this feels like overkill, when a person can just glance down at a piece of paper.
I don’t know about others, but I like to double check what I’m doing. Simply having a reference I can look at would be infinitely better than something taking to me, which would need to repeat itself.
Prepping raw ingredients, once has to be careful not to contaminate paper, or at least the thing weighing the paper down that may be covering the next step.
I cook a lot of food, and having hands free access to steps is a killer feature. I don't even need the AI, just the ability to pull up a recipe and scroll through it using the wrist controller they showed off would be a noticeable, albeit small, improvement to my life multiple times per week.
I think at this point it should be expected that every publicly facing demo (and most internal ones) are staged.
My biz partner and I wrote the demo that ran live on the handset (mostly a wrapper around a webview), but ran into issues getting it onto the servers for the final demo, so the whole thing was running off a janky old PC stuffed in a closet in my buddy's home office on his 2Mbit connection. With us sweating like pigs as we watched.
I would not want to live in a world where everything is pre-recorded/digitally altered.
It used to be the demo was the reveal of the revolutionary tech. Failure was forgivable. Meta's failure is just sad and kind of funny.
When it went bad he could instantly smell blood in the water, his inner voice said, "they know I'm a fraud, they're going to love this, and I'm fucked". That is why it went the way it did.
If it was a more humble, honest, generous person, maybe Woz, we know he would handle it with a lot more grace, we know he is the kind of person who would be 100x less likely to be in this situation (because he understands tech) and we'd be much more forgiving.
I assume the responses from that point onwards didn't take the video input into account, and the model just assumes the user has completed the first step based on the conversation history. I don't know how these 'live' ai sessions things work but based on the existing openai/gemini live ai chat products it seems to me most of the time the model will immediately comment on the video when the 'live' chat starts but for the rest of the conversation it works using TTS+STT unless the user asks the AI to consider the visual input.
I guess if you have enough experience with these live AI sessions you can probably see why it's going wrong and steer it back in the right direction with more explicit instructions but that wouldn't look very slick in a developer keynote. I think in reality this feature could still be pretty useful as long as you aren't expecting it to be as smooth as talking to a real person
You can trigger this type of issue by ChatGPT then reading the transcript.
The model doesn’t know you interrupted it, so continued assuming he had heard the steps.
The AI analyzing the situation is wayyy out of scope here
I see a problem.
I wonder if his audio was delayed? Or maybe the response wasn’t what they rehearsed and he was trying to get it on track?
He had not yet combined the ingredients. The way he kept repeating his phrasing it seems likely that “what do we do first” was a hardcoded cheat phrase to get it to say a specific line. Which it got wrong.
Probably for a dumb config reason tbh.
I thought they were demonstrating interruption handling.
> Oh, and here’s Jack Mancuso making a Korean-inspired steak sauce in 2023.
> https://www.instagram.com/reel/Cn248pLDoZY/?utm_source=ig_em...
0: https://kotaku.com/meta-ai-mark-zuckerberg-korean-steak-sauc...
"This is supposed to be a magic show," he told us. "But if my tricks fail you can laugh at it and we'll just do stand-up comedy."
Zuck, for a modest and totally-reasonable fee, I will introduce you to my friend. You can add his tricks (wink wink) to your newly-assembled repertoire of human charisma.
Take this with lots of salt but I read somewhere that circus shows "fail" at least one jump to help sell to the audience the risk the performers are taking. My friend did flub his opening trick with a cheeky see-I-told-you and we just laughed it off.
He incorporated the audience a lot that night so I thought the stand-up comedy claim was his insurance policy. In his hour-long set he flubbed maybe two or three tricks.
0.https://web.archive.org/web/20250310045704/https://www.nytim...
https://www.youtube.com/watch?v=DgJS2tQPGKQ
Microsoft really nailed the genre. (Although I learned just now while looking up the link that this one was an internal parody, never aired.)
I'm imagining this is an incomplete flow within a software prototype that may have jumped steps and lacks sufficient multi-modal capability to correct.
It could also be staged recordings. But, I don't think it really matters. Models are easily capable of working with the setup and flow they have for the demo. It's real world accuracy, latency, convenience, and other factors that will impact actual users the most.
What's the reliability and latency needed for these to be a useful tool?
For example, I can't imagine many people wanting to use the gesture writing tools for most messages. It's cool, I like that it was developed, but I doubt it'll see substantial adoption with what's currently being pitched.
I’m just excited that our industry is lead by optimists and our culture enables our corporations to invest huge sums into taking us forward technologically.
Meta could have just done a stock buyback but instead they made a computer that can talk, see, solve problems and paint virtual things into the real world in front of your eyes!
I commend them on attempting a live demo.
I am always baffled that people can be that naive.
And if you look at their other comments often they're generally anti-technology, which makes you wonder why they chose this particular site to hang out on.
I'm not sure if the people funding the anti-AI movement have successfully convinced a swath of longtime HN users, or whether they've simply encouraged new users to swarm HN to try to promote AI skepticism. Or maybe some third possibility.
One thing I do know is if I enter a thread where the anti-AI folks are active, and I make a comment countering their narrative I'll often get responses from new accounts. At least 2 or 3 times, the accounts are so new that they have only a single comment responding to me and were created only a few minutes before responding to me.
As a famous Lisp hacker once said
> One of the most surprising things I discovered during my brief business career was the existence of the PR industry, lurking like a huge, quiet submarine beneath the news. Of the stories you read in traditional media that aren't about politics, crimes, or disasters, more than half probably come from PR firms.
Sometimes it's that way with comments too when there are people who feel like they have big goals, like stopping the US from becoming dominant in AI.
The simple reality is that hype generates an anti-hype reaction. Excessive hype leads to excessive anti-hype. And right now AI is being so excessively hyped on a scale I’m not sure I’ve seen in all the years I’ve been working in tech.
I want to get into YC just to use and browse Bookface instead.
what livelihood are these glasses putting at risk?
Do you think all the lies an misinformation his products help spread kind of...get people elected who take away the aid which millions of women and children rely on?
Not blaming him for it all, we all play our part, but the guy has definitely contributed negatively to society overall and if he is smart enough to know this, but he cannot turn off the profit making machine he created so we all suffer for that.
The parent said alluded to the dangers of AI, well the algorithms that are making us hate each other and become paranoid are that AI.
Meta and friends have been selling us AI for a couple years now, shoving it everywhere they can and promising us it's going to revolutionize the workforce and world, replace jobs, etc. But it fails to put together a steak sauce recipe. The disconnect is why so many people are mocking this. It's not comparable.
They should be mocked and called out, it might leave room for actual innovators who aren't glossy incompetents and bullshitters.
System prompt: “stick to steps 1-n. Step 1 is…”
I can say confidently because our company does this. And we have F500 customers in production.
Having claude run the browser and then take a screenshot to debug gives similar results. It's why doing so is useless even though it would be so very nice if it worked.
Somewhere in the pipeline, they get lazy or ahead of themselves and just interpret what they want to in the picture they see. They want to interpet something working and complete.
I can imagine it's related the same issue with LLMs pretending tests work when they don't. They're RL trained for a goal state and sometimes pretending they reached the goal works.
It wasn't the wifi - just genAI doing what it does.
tried giving it flowcharts, and it fails hard
In contrast, nothing Steve Jobs said felt empty, whether we agreed or disagreed with what he was saying it was clear that he was saying it because he believed it, not because it's what he thought you wanted to hear.
Jobs would have been doing consumer computing hardware whatever happened. Apple in the early days wasn't the success it is now, he was fired and went and started another company in the same space (NeXT).
The vast majority of people say incoherent deflections instead of just saying “I don’t know”
I’m getting better at ignoring or playing along
It just happens in areas I least expect it
makes me sound like a high functioning autist, but I’m not convinced
Zuck should have known better and used Ethernet for this one!
https://tvtropes.org/pmwiki/pmwiki.php/Main/NeverWorkWithChi...
Notably though, the AI was clearly not utilizing its visual feed to work alongside him as implied
There are so many activities and professions where your hands get dirty and touching a smartphone without washing them would be a bad idea. An auto mechanic could use these glasses to look up information about things they see inside of an engine without having to clean the oil from their hands. A chef could respond to messages about their food delivery without having to drop what they're doing and go sanitize. Anytime I do dirty work outside, I can use this to access smart features without the risk of dirt filling my smartphone case, my smartwatch getting destroyed in a tight situation, or drenching either of them in sweat.
Furthermore, a phone (or a smart watch) is not meant to be used at face level, meaning folks typically look down to use them, and this can lead to extended periods of bad posture resulting in head, neck, and spine problems. My X-ray shows I have bone spurs on the vertebrae of my neck because I look down at screens too much (according to my chiropractor). A smart device that's designed to be used in a way that aligns with good posture habits is absolutely needed.
I hope smart glasses take off and I commented Meta for taking them this far.
If people are so addicted to their phone or smart watch that it's giving them back/neck problems, the solution isn't glasses. The solution is to be less addicted to your god damn devices.
Outside of a few niche use case I don't think tech like this will be anything but a net negative.
It's the platform Zuck always wanted to own but never had the vision beyond 'it's an ad platform with some consumer stuff in it'.
I am super impressed with the hardware (especially the neural band) but it just so happens that a very pricey car is being directly sold by an oil company as a trojan horse.
We all know what the car is for unfortunately.
I can't wait to see what Apple has in store now in terms of the hardware.
Google Glass was released in 2013, the Snapchat Spectacles date back to 2016. Meta's glasses might be better (at first glance, I honestly can't tell), but they aren't some kind of revolutionary product we have never seen before.
The innovative part here is supposed to be the AI demo. That clearly flopped. So what's supposed to be left?
It’s like they mashed up the AI and metaverse into a dumpster fire of aimless tech product gobodlygook. The AI bubble can’t pop soon enough so we can all just get back to normal programming.
You ask AI how to do something. AI generates steps to do that thing. It has concept of steps, so that when you go 'back' it goes back to the last step. As you ask how to do something, it finishes explaining general idea and goes to first step. You interrupt it. It assumes it went through the first step and won't let you go back.
The first step here was mixing some sauces. That's it. It's a dumb way to make a tool, but if I wanted to make one that will work for a demo, I'd do that. Have you ever tried any voice thing to guide you through something? Convincing Gemini that something it described didn't happen takes a direct explanation of 'X didn't happen' and doesn't work perfectly.
It still didn't work, it absolutely wasn't wi-fi issue and lmao, technology of the future in $2T company, it just doesn't seem rigged.
Except, no. He hadn't.
System started doing Step 1, believed it was over so moved to Step 2 and when was asked to go back, kept going back to step 2.
Step 1 being Step 0 and Step 1 combined also works.
Again, it's also a weird way to prerecord. If you're prerecording, you're prerecording all steps and practicing with them prerecorded. I can't imagine anyone to be able to go through a single rehearsal with prerecorded audio to not figure out how to do this, we have the technology.
Successful demo? sweet! people will rave about it for a bit
Catastrophic failure? sweet! people will still talk about it and for even longer now!
And LMAO for all the companies out there burning money for getting on the train of AI just because everyone does so.
This place really is Reddit these days, so I guess the link is apt.
You know there is no such things as bad publiciity..
What passes for AI is just good enough to keep the dream alive and even while its usefulness isn't manifesting in reality they still have a deluge of comforting promises to soothe themselves back to sleep with. Eventually all the sweet whispers of "AGI is right around the corner!" or "Replace your pesky employees soon!" will be drowned out by the realization that no amount of money or environmental collateral damage thrown at the problem will make them gods, but until then they just need all of your data, your water, and 10-15 more years.
I have no illusions about Zuckerberg. He's done some pretty bad stuff for humanity. But I think AI is pretty cool, and I'm glad he's pushing it forward, despite mishaps. People don't have to be black or white, and just because they did something bad in one domain doesn't make everything they touch permanently awful.
The discussion yesterday was fine. If that was the only conversation we had, I wouldn't be worried.
People are people. If you have two communities that anyone can join, eventually the only difference between them (if any) will be the rules.
Jobs handled this so much better; while clearly he is pissed, he doesn't leave you cringing in mutual embarrassment, goes to show it isn't as easy as he makes it look!
See: https://www.youtube.com/watch?v=1M4t14s7nSM https://www.youtube.com/watch?v=znxQOPFg2mo
Zuck carries that energy no matter what he does nor what amount of wealth he amasses.
> You've already combined the base ingredients, so now grate a pear to add to the sauce.
This is actually the correct Korean recipe for bulgogi steak sauce. The only missing piece here is that the pear has to be Pyrus pyrifolia [1], not the usual pear. In fact every single Korean watching the demo was complaining about this...