u/Tall-Peak2618

A few weeks running an end to end VLA on a real arm and some things I did not expect

Been quietly swapping our usual perception/planning/control stack for an end to end VLA model on a UR style arm + parallel gripper setup. Mostly because my advisor wanted to see if the hype was real, and because two of the open weights releases this spring (pi0.6 and the WALL OSS drop from X Square Robot) actually run on a single 4090 without too much pain.

Some stuff that genuinely caught me off guard, in no particular order.

The good. Recovery behavior is weirdly fluent. With our old stack, if the grasp slipped we hit a planning re-call and the arm would just stop for ~400ms and then redo the whole motion. The VLA just adjusts mid trajectory the way a person would, it doesnt look like a state machine recovering, it looks like a hand. I have no good explanation for why this is the part that surprised me most, but it is.

The annoying. Latency variance is awful at the start. First few hundred episodes of fine tuning, we were seeing 80 to 240 ms inference jitter on the same hardware. Turns out a lot of that was us still feeding it preprocessed depth from our old pipeline, which the model didnt want. Once we just gave it raw RGB and proprio it stabilized.

The unexpected. Language conditioning is not magic. "pick up the red one" works. "pick up the red one and put it on the cloth, not the plate" is a coin flip in our setup. Multi clause instructions still fall apart in ways that feel very 2022. I think people see the demos and assume natural langauge is solved, it is very much not, at least not at our scale.

The philosophical one. After a while it becomes hard to tell what the model is "doing wrong". With a modular stack, when something fails you can point at it: localization drifted, the planner chose a bad pose, the controller overshot. With end to end you just get a worse rollout and a vague feeling. The interpretability story for VLAs is going to be a real problem for anyone shipping this in safety critical contexts.

Not selling anything, not affiliated with the labs releasing these weights. Honestly the main reason I am writing this up is because all the public discourse is either "lab demo of the century" or "it is all teleop", and the actual day to day experience of running one of these things is much more boring and much more interesting than either.

If you have run pi0.6, WALL OSS, OpenVLA or anything in that family on real hardware (not sim), drop your weirdest observation. I will collect them and post a follow up if there is enough material.

reddit.com
u/Tall-Peak2618 — 3 days ago
▲ 2 r/SaaS

Built a game studio with ai. 20m users in 8 months, $0 in dev costs.

Li Fei-Fei's ai game company astrocade just raised $56m. 20m registered users, 5m mau, 1.4b monthly plays. top creators making thousands a month.

Platform lets anyone build games by describing them in natural language. no code, no engine. just text prompts and a few minutes of waiting.

Been testing it the past week. heres what the builder side actually looks like.

Workflow is basically: describe your game in plain language ("space shooter with asteroids and power ups"), wait 5 to 10 minutes, get a playable game with core mechanics, basic assets, ui. then iterate through chat ("add a shield power up", "make enemies faster"). publish to the feed.

Stack under the hood is terrain models, character animation models, an astrobrain coordination model that orchestrates everything, plus an editor for fine tuning colors, speed, difficulty.

What works: simple mechanics, shooters, puzzles, basic platformers. iteration is fast once the base game exists. tiktok style feed drives discovery. monetization is built in.

What doesnt: complex games get confused. tried a strategy game, got a mess. asset quality is inconsistent. youre locked into their platform.

The one person company angle is what im actually thinking about. infra layer is mature enough that one person plus ai tools can build what used to need teams. games on astrocade. code via platforms like verdent that handle multi agent orchestration. content via claude or gpt. design from midjourney or figma ai.

Bottleneck isnt technical anymore. its distribution, product judgment, knowing what to build.

Astrocades 20m came from the platform itself. discovery engine is the product. you dont need marketing, you need a game thats good enough to stop people scrolling.

Top creators making 3-5k a month. not life changing but real revenue from games that took hours to build.

Counter argument is platform risk. building on rented land. astrocade owns distribution, takes their cut, can change terms. trade off for solo builders is clear though, speed vs ownership.

reddit.com
u/Tall-Peak2618 — 5 days ago

Non technical HR here, tried 4 agent platforms for candidate research and one finally fit

Posting because most of what i read on this sub is for developers and i wanted to share what worked for someone who genuinely cant read code.

Im in HR at a 22 person company. We hire 3 to 5 roles a month, mostly through referrals and one job board. For each candidate that makes it past the first screen i need a deeper look, what they actually built, who they actually worked with, whether their writing matches what their resume claims. Used to spend 30 to 45 min on each one going through linkedin, their portfolio, project links, sometimes their twitter. When the week got busy i just skipped that step and i could feel it in the interviews.

Last month i tested 4 agent platforms after a friend in marketing told me they were not just for people who can code. Mixed results.

Bardeen. Tried this first because the chrome extension installs in 30 seconds. For a top of profile summary it works out of the box, no setup really. Where it stopped was when i wanted depth, projects, recommendations, who endorsed what at that level. Could not get past the surface profile without help i didnt have. Also hit some linkedin limits when i tried to run too many profiles in a row. Good for quick summaries, not for the deep dive i needed.

Apollo.io. We already use this for sales. I know its not really meant for candidate research but i tried anyway. The data was flat. Static profile, no recent activity, no project links. Confirmed its better as a contact db than a research tool. Kept it for what its good at.

Manus. Wrote me a clean summary of each candidate but pulled some of the details from old web mentions that werent the right person. Same name, different person. Got embarrassing once when i quoted something in an interview that turned out to be from someone elses portfolio. Trust issue after that. If you use this, double check every claim against the actual source.

MuleRun. The part that actually makes this work for me is the browser extension. I open the candidates linkedin in chrome, click the extension button, and it does its thing. Scrolls through their profile, opens their listed projects in new tabs, opens any cited articles, and about 8 min later i have a 1 page summary saved to a candidate folder in drive.

Before this each candidate took me 30 to 45 min and i often skipped the deep context because of time pressure. Now i review in 5 to 8 min and i dont skip. One honest weakness, the very first time i set it up took about 20 min because i had to teach it which fields i actually care about (years per role, listed projects, whether the candidate has any public writing, references to specific tools they list). After that it stayed consistent across 80 plus candidates. So the first run feels slow. After that its fine.

Ended up keeping mulerun for the deep candidate research, apollo for the contact side. Bardeen and manus i let go.

Couple things i wish someone had told me before i started. The biggest unlock for non technical people is anything that runs in your already logged in browser. I do not want to share my linkedin password with a tool. I do not want to set up an api. The extension model just works. Also the first hour of setup feels worse than the manual process you are replacing. Push through. The 80th candidate is what you optimize for, not the 1st.

What i havent solved yet, where these briefs go after a hire is finalized. They just sit in their drive folder right now. Will probably wire them into our ATS down the road.

reddit.com
u/Tall-Peak2618 — 6 days ago

Not some guru discovery story. Was just messing around trying to see if AI could do anything useful for product research beyond writing listing copy.

Background on me: been selling on amazon for about 2 years, mostly in home and kitchen. Decent revenue but growth flatlined because every time i find something promising, there's already 40 other sellers sitting on it.

About 6 weeks ago i set up a weekly scan in MuleRun to track category level search data and flag anything where searches are climbing but seller count isn't keeping up. Honestly didn't expect much from it. Ran for maybe a month in the background while i focused on other stuff.

Then it flagged a sub-category i'd never even thought to look at. Search volume up roughly 3x over 90 days. Active seller count barely moved. Top listings had mediocre reviews, 3.5 to 3.8 average, and the price points left room for a mid-range entry.

Won't name the exact niche since i'm currently sourcing samples, but its adjacent to outdoor/seasonal stuff. The point isn't the specific product anyway. I spent maybe 15 minutes a week on this and it surfaced something i would've never noticed scrolling through amazon manually.

My old process was jungle scout plus manually checking BSR movement plus reading review complaints. Still works but it's slow and you only cover categories you already know about. Having something scan across categories you wouldn't think to check yourself is different.

Biggest takeaway for me was that the opportunity wasn't in a category i was already watching. It was in a completely adjacent space i wouldn't have bothered looking at on my own. That's where the automated scanning actually adds value vs just doing what i was already doing faster.

reddit.com
u/Tall-Peak2618 — 8 days ago
▲ 3 r/Home

Recently i stepped away from work for a bit, mostly just needed some space from the usual rush. I set up a small metal gazebo in the yard and added mesh around it so bugs wouldn't bother me. It's one of those costway ones, fit the spot perfectly. After that, i found myself going out there more often.

Some afternoons i just sit, sometimes lie down for a bit, let my mind wander and the breeze pass by. After a few days i noticed little things i hadn't before—the way sunlight hits certain corners, plants that could use a bit of rearranging. So i moved a few pots around, added a couple new ones, and slowly the area started to feel more alive. It didn't solve everything, but having that little spot gives me a part of the day that feels calm and steady. And honestly, that's been enough lately.

reddit.com
u/Tall-Peak2618 — 8 days ago

I’m thinking about getting this brown off shoulder top, but I’m not sure how I would style it. Would it work better with pants or a skirt for a casual dinner? I was especially wondering if a white skirt would look good with this color, or if another pairing would be better.

u/Tall-Peak2618 — 14 days ago
▲ 1 r/mcp

I've been following the discussions here about the actual practical value of MCPs versus standard REST APIs. After spending the last few weeks building, I genuinely believe crypto/Web3 data is the killer use case for the MCP architecture.

Here is the problem: Crypto data is insanely fragmented. If you want to do deep project research, you're pulling on-chain metrics, CEX/DEX real-time pricing, social sentiment, and protocol fundamentals from dozens of different platforms. Standard API aggregation is a nightmare, especially for non-devs.

MCP solves this perfectly because the LLM can just dynamically route and pull exactly the context it needs without you writing custom API wrappers for every single source.

To test this, I built Surf (https://usesurf.com) — a zero-code MCP data skill layer specifically for deep token and project research. It lets Claude or any MCP-compatible LLM directly query this fragmented data.

Instead of writing scripts, you can just prompt your local agent to:

  • Automatically compare the TVL (Total Value Locked) trends of two different DeFi protocols over the last 30 days.
  • Query the holder concentration and recent whale movements for a specific token in one sentence.
  • Cross-reference real-time market cap with underlying protocol revenue.

I'm handling the API routing, rate limits, and data normalization on the backend so the agent just gets clean context to work with.

I'm curious to hear from other builders — outside of crypto, what other highly fragmented data verticals do you think are ripe for dedicated MCP data skills? Traditional finance? Real estate?

reddit.com
u/Tall-Peak2618 — 16 days ago
▲ 97 r/dji

Frame is packed with greens, dark tones up close fading to lighter yellowish green on distant ridges, all smooth with no banding. I've shot vegetation on overcast days before and it usually looks flat and fake, but here the colors held up even without direct sun doing the work.

Around 6 to 18s the fog rolls in and swallows more than half the mountain. That treeline to mist boundary is barely there contrast wise and most sensors would just mush it together, but can still make out layers even at its thickest.Better than I expected going in.

u/Tall-Peak2618 — 18 days ago

I do freelance audio engineering and I've had mild hearing loss in my left ear since my twenties (too many years standing next to guitar amps). When Apple rolled out the hearing aid feature I was genuinely excited to try it, but I kept running into this thing where my own voice sounded hollow and doubled when I talked. Not a dealbreaker for listening to other people, but really disorienting during conversations because I could feel myself starting to speak more quietly or weirdly just to avoid hearing that strange echo.

So I went down a rabbit hole trying to understand what was happening from a signal processing perspective, and it turns out this is a well documented phenomenon. When you wear any device that processes sound and plays it back into your ear canal, there are two versions of your voice reaching your cochlea. The first arrives almost instantly through bone conduction, the way you always hear yourself. The second arrives after the device captures it with a microphone, runs it through whatever DSP chain it uses, and outputs the processed version. If the gap between those two arrivals is large enough, your brain perceives them as separate events. You get a comb filtering effect where certain frequencies reinforce and others cancel, and the result is that hollow, phasy, "talking inside a tin can" sensation.

The critical threshold most psychoacoustic research points to is around 10 milliseconds. Below 10ms, the brain fuses the two signals into one perceived sound. Above 10ms, you start to consciously detect the delay as a distinct second arrival. It does not have to be a dramatic echo. Even at 12 or 15ms you just feel something is off. Your voice sounds thicker in a bad way, or you get a subtle flange that makes you want to stop talking.

The challenge for AirPods Pro 2 in hearing aid mode is structural. The audio path has to go through the Bluetooth codec layer and then through the transparency mode DSP pipeline, and both of those add latency. Apple has done incredible work getting transparency mode latency down for a consumer earbud, but the architecture was designed for a different primary use case. The hearing aid function is layered on top of hardware that was optimized for music playback and active noise cancellation first. That is not a criticism, it is just a design priority reality.

I ended up testing an OTC hearing aid built around a dedicated hearing aid chipset, specifically the ELEHEAR Beyond Pro, which specs its processing latency at 8ms or less. The difference when speaking was immediately obvious. That doubled, phasy quality I kept hearing with the AirPods transparency pipeline was basically gone. My voice just sounded like my voice, which honestly I had started to think was an unsolvable problem with any in ear amplification device. It was not. It was a latency problem.

What made it click for me technically is that dedicated hearing aid DSP chips are designed from the ground up with a single architectural priority: minimize the time between microphone input and speaker output. Every gate in the signal path is optimized for that. Consumer audio chips are balancing a dozen other priorities like codec flexibility, ANC computation, spatial audio rendering, and latency is one constraint among many rather than the dominant one.

None of this means AirPods are bad as hearing aids. For a lot of people they work great, and the accessibility win of turning a device millions already own into a hearing aid is genuinely important. But if you have been bothered by how your own voice sounds and assumed that is just what wearing hearing aids feels like, it might actually be a latency artifact that a sub 10ms device resolves completely.

reddit.com
u/Tall-Peak2618 — 29 days ago