Hacker Newsnew | past | comments | ask | show | jobs | submit | silverlake's commentslogin

I used ET but it requires a server process also. Some machines are too locked down to allow this. Wish there was a way to kick start the server on demand.


Uh, mosh needs to be installed to the server as well?


Well yes, but mosh starts its server over an initial SSH connection used for setup, so you only need the binary to exist in PATH of the remote host and you're done. It's more difficult to arrange a service to be running; sometimes more so if you don't have root.


I’m actually working on just this. What’s the smallest training data set required to learn tic-tac-toe? A 5yo doesn’t need much training to learn a new game, but a transformer needs millions of samples.


> A 5yo doesn’t need much training to learn a new game

A 5yo also has... 5 years of cumulative real world training. I'm a bit of an AI naysayer but I'd say the comparison doesn't seem quite accurate.


It’s a glib analogy, but the goal remains the same. Today’s training sets are immense. Is there an architecture that can learn something with tiny training sets?


Maybe ZephApp, when it's actually released. But would be interesting to record day-to-day conversations (face-to-face using voice recognition) to train a virtual doppelganger of myself and use it to find uncommon commonalities between myself and others.

What would someone do with a year's worth of recorded conversations? Would the other parties be identified? How would it be useful, if at all? How about analyzing the sounds/waveform rather than words? (eg BioAcousticHealth / vocal biomarkers)

Perhaps typing into a text-field is the problem right now? Maybe have a HUD in a pair of glasses. Better than getting a brain chip! Most recent or most repeated conversations most important. Could lead to a reduction in isolation within societies, in favor for "AI training parties." Hidden questions in oneself answered by a robot guru as bedtime story-telling but related to the real-world and real-events.

Smart Glasses --> Smart Asses

Vibe Coding --> Tribe Loading

Everything Probable --> Mission Impossible


I'm certainly not challenging anything you're writing, because I only have a very distant understanding of deep learning, but I do find the question interesting.

Isn't there a bit of a defining line between something like tic-tac-toe that has a finite (and pretty limited for a computer) set of possible combinations where it seems like you shouldn't need a training set that is larger than said set of possible combinations, and something more open-ended where the impact of the size of your training set mainly impacts accuracy?


Assuming you don't account for reflections, rotations, and 'unreachable' gamestates where a player wins and you continue to mark boxes.

It's just 3^9, right? 9 boxes, either X,O, or blank? We're only at 19,683 game states and would trim down from here if we account for the cases above.


Exactly, but then we may as well say "don't solve this with an LLM" which sort of kills the conversation altogether and that's not my goal. :)


Oh, im sorry! I was just trying to give a quick perspective of how small that tic-tac-toe data-set actually is. Not suggest against the idea!


Oh no worries at all. :)


And hundreds of millions of years of evolutionary intelligence.


Next step in AI: teaching an LLM to think like a trilobite!


A trilobite was obviously better at being a trilobite than an LLM would be, if not by purely definitional purposes.


Was the six million dollar man not a better man?


This sounds super interesting. Will you be sharing your work anywhere? :)


Also, most of them are shockingly incompetent. It took years to assemble a list of quality service providers. I pay a little more but stuff works now.


There is a mountain of code that needs to be written that can’t be due to costs. A project that needs 100 developers may be prohibitively expensive. But 10 10x developers would be within budget. Think health care, manufacturing, government, finance, etc.


There is also a mountain of code that shouldn't have been written, and increasingly it's Ai generated.


I tried to get Gemini CLI to update itself using the MCP settings for Claude. It went off the rails. I then fed it the link you provided and it correctly updates it's settings file. You might mention the settings.json file in the README.


I’ve been talking to wealth management firms and am truly underwhelmed. AFAICT, they charge 0.5%-1.35% for therapy and the chance to put you in high fee products. Roboadvisors are a brilliant product for most people. In fact, a simple Boglehead portfolio is all you need. Most people have simple and similar risk profiles.

The feature I think would be useful is how to manage taxes. Roth conversions, selling the right lot, qualified dividends, tax loss harvesting, etc. A related feature would be generating income while minimizing taxes, i.e. Schwab’s Intelligent Income.


Betterment's roboadvisor has tax loss harvesting. It does the basics like selling the right lot; I'm not educated enough about finance to say what it's still missing (that would be easy for a bot to do).

https://www.betterment.com/tax-loss-harvesting

https://www.betterment.com/resources/tax-loss-harvesting-met...


You’re clinging to an old model of work. Today an LLM converted my docker compose infrastructure to Kubernetes, using operators and helm charts as needed. It did in 10 minutes what would take me several days to learn and cobble together a bad solution. I review every small update and correct it when needed. It is so much more productive. I’m driving a tractor while you are pulling an ox cart.


“ It did in 10 minutes what would take me several days to learn and cobble together a bad solution.”

Another way to look at this is you’re outsourcing your understanding to something that ultimately doesn’t think.

This means 2 things: your solution could be severely suboptimal in multiple areas such as security and two because you didn’t bother understanding it yourself you’ll never be able to identify that.

You might think “that’s fine, the LLM can fix it”. The issue with that is when you don’t know enough to know something needs to be fixed.

So maybe instead of carts and oxen this is more akin to grandpa taking his computer to Best Buy to have them fix it for him?


Senior engineers delegate to junior engineers, which have all the same downsides you described, all the time. This pattern seems to work fine for virtually every software company in existence.


> Another way to look at this is you’re outsourcing your understanding to something that ultimately doesn’t think.

You read this quote wrong. Senior devs outsource _work_ to junior engineers, not _understanding_. The way they became senior in the first place is by not outsourcing work so they could develop their understanding.


How about a CEO delegating the work to an Engineer ? CEO does not understand all the technical detail but only knows what the outcome will look like.


I read the quote just fine. I don't understand 100% of what my junior engineers do. I understand a good chunk, like 90-95% of it, but am I really going to spend 30 minutes trying to understand why that particular CSS hack only works with `rem` and not `px`? Of course not - if I did that for every line of code, I'd never get anything done.


You are moving goalposts significantly here -- a small CSS hack is a far cry from your docker infrastructure.


I am going to put it out here: Docker and other modern infra is easier to understand than CSS (at least pre flex).


My take from this comment is that maybe you do not understand it as well as you think you do. Claiming that "other modern infrastructure" is easier to understand than CSS is wild to me. Infrastructure includes networking and several protocol, authentication and security in many ways, physical or virtual resources and their respective capabilities, etc etc etc. In what world is all of that more easy than understanding CSS?


When did I say I was blindly allowing an AI to set up my docker infrastructure? Obviously I wouldn't delegate that to a junior. My goalposts have always been in the same place - perhaps you're confusing them with someone else's goalposts.


I have been coding 10+ years, surely it is fine for me to vibecode then?


Only if you don’t mind what comes out :)


I mean I love it.


Comparing apples to oranges in your response but I’ll address it anyway.

I see this take brought up quite a bit and it’s honestly just plain wrong.

For starters Junior engineers can be held accountable. What we see currently is people leaving gaping holes in software and then pointing at the LLM which is an unthinking tool. Not the same.

Juniors can and should be taught as that is what causes them to progress not only in SD but also gets them familiar with your code base. Unless your company is a CRUD printer you need that.

More closely to the issue at hand this is assuming the “senior” dev isn’t just using an LLM as well and doesn’t know enough to critique the output. I can tell you that juniors aren’t the ones making glaring mistakes in terms of security when I get a call.

So, no, not the same. The argument is that you need enough knowledge of the subject call bs to effectively use these tools.


> For starters Junior engineers can be held accountable. What we see currently is people leaving gaping holes in software and then pointing at the LLM which is an unthinking tool. Not the same.

This is no different than, say, the typical anecdote of a junior engineer dropping the database. Should the junior be held accountable? Of course not - it's the senior's fault for allowing that to happen at the first place. If the junior is held accountable, that would more be an indication of poor software engineering practices.

> More closely to the issue at hand this is assuming the “senior” dev isn’t just using an LLM as well and doesn’t know enough to critique the output.

This seems to miss the point of the analogy. A senior delegating to a junior is akin to me delegating to an LLM. Seniors have delegated to juniors long before LLMs were a twinkle in Karpathy's eye.


> This is no different than, say, the typical anecdote of a junior engineer dropping the database. Should the junior be held accountable? Of course not - it's the senior's fault for allowing that to happen at the first place. If the junior is held accountable, that would more be an indication of poor software engineering practices.

Of course the junior should be held accountable, along with the senior. Without accountability, what incentive do they have to not continue to fuck up?

Dropping the database is an extreme example because it's pretty easy to put in checks that should make that impossible. But plenty of times I've seen juniors introduce avoidable bugs simply because they did not bother to test their code -- that is where teaching accountability is a vital part of growth as an engineer.


The second part of my response addresses why your response isn’t analogous to what we’re discussing.


No one is an expert on all the things. I use libraries and tools to take care of things that are less important. I use my brain for things that are important. LLMs are another tool, more flexible and capable than any other. So yes, grandpa goes to Best Buy because he’s running his legal practice and doesn’t need to be an expert on computers.


True, but I bet grandpa knows enough to identify when a paralegal has made a case losing mistake ;)


I am pretty confident that my learnings have massively sped up working together with LLMs. I can build so much more and learn through what they are putting out. This goes to so many domains in my life now, it is like I have this super mentor. It is DIY house things, smart home things, hardware, things I never would have been confident to work with otherwise. I feel like I have been massively empowered and all of this is so exciting. Maybe I missed a mentor type of guidance when I was younger to be able to do all DYI stuff, but it is definitely sufficient now. Life feels amazing thanks to it honestly.


If there's something that you don't understand, ask the LLM to explain it to you. Drill into the parts that don't make sense to you. Ask for references. One of the big advantages of LLMs over, say, reading a tutorial on the web is that you can have this conversation.


> I’m driving a tractor while you are pulling an ox cart.

Or you’re assembling prefab plywood homes while they’re building marble mansions. It’s easy to pick metaphors that fit your preferred narrative :)


>you’re assembling prefab plywood homes while they’re building marble mansions

Which one are there more of nowadays, hm?


Maybe the least interesting question to ask. Instead: Which ones are more lucrative to work on? Which ones are more fun to work on?


> would take me several days to learn ... correct it when needed.

If you haven't learned how all this stuff works, how are you able to be confident in your corrections?

> I’m driving a tractor while you are pulling an ox cart.

Are you sure you haven't just duct taped a jet engine to your ox cart?


How did you verify this works correctly, and as intended, in 10 minutes if it would have taken you 2 days to do it yourself?


But you would have learned something if you invested the time. Now when your infra blows up you have no idea what to fix and will go fishing into the LLM lake to find how to fix it


> It did in 10 minutes what would take me several days to learn

> I review every small update and correct it when needed

How can you review something that you don't know? How do you know this is the right/correct result beyond "it looks like it works"?


If it would have taken you days to learn about the topic well enough to write a bad implementation, how can you have any confidence you can evaluate, let alone "correct", one written by an LLM?

You just hope you are on a tractor.



Here’s the real rebuttal to my overconfidence in LLMs. Thanks for the link!


I think this fits squarely with the idea that LLM today is a great learning tool; learning through practice has always been a proven way to learn but a difficult method to learn from fixed material like books.

LLM is a teacher that can help you learn by doing the work you want to be doing and not some fake exercise.

The more you learn though, the more you review the code produced by the LLM and the more you'll notice that you are still able to reason better than an LLM and after your familiarity with an area exceeds the capabilities of the LLM the interaction with the LLM will bring diminishing returns and possibly the cost of babysitting that eager junior developer assistant may become larger than the benefits.

But that's not a problem, for all areas you master there will be hundreds of other areas you haven't mastered yet or ever will and for those things the LLM we have already today are of immediate help.

All this without even having to enter the topic of how coding assistants will improve in the future.

TL;DR

Use a tool when it helps. Don't use it when it doesn't. It pays to learn to use a tool so you know when it helps and when it doesn't. Just like every other tool


This is such an arrogant take.


Have you guys at Cline considered using LLMs to create summaries of files and complex functions? Rather than read a 500 line function, feed it a short comment on what the function is doing. I'd like to use a local LLM to create summaries at every level: function, file, directory. Then let the LLM use that to find the right code to read. This is basically how I navigate a large code base.


I've just used Cline to produce files like that, and then later when starting a task in plan mode I tell it to read those files to get a sense of the project's structure. I also tell it to update them as necessary after whatever task we're doing is finished.


So you're effectively keeping two copies of the codebase but the second copy is written in prose?


It's basically an index.


aka documentation


Would it double your codebase? Do you think it would work for a large codebase?


Not anywhere close. It's basically just maintaining a simple descriptive index the model can later use to decide what files it needs to read given the task you've given it.


No offense, but that video is brutally boring. Even at 1.5x speed I couldn’t get past 10 min. You should transcribe the audio and use an LLM to write a punchy sales pitch.


If you look at the full table of NAEP scores, you see the differences are not that much. 8th grade reading is 245(WV)-266(MA). That's 20 point spread, where NAEP goes to 500 pts. A 4% difference between the best and worst state is insignificant. They all suck because they are at 50% of the scale.


From TFA - a 10 point gap in NAEP represents approximately one grade level, so your 21 point spread is the difference between 3rd and 5th grades.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: