Hacker News .hnnew | past | comments | ask | show | jobs | submit | ccleve's commentslogin


But the author of that page is not concerned with readability or accessibility. He just wants things to look cool and design-y. One piece of supporting evidence he cites is some random photo he took that doesn't contain #000000 black. That doesn't mean anything, it could be that it's over-exposed, or has poor contrast, or had some silly filter applied. This leads me to think that the author of that page doesn't know what he's talking about.


Additions without any evidence:

> He just wants things to look cool and design-y

> some random photo

> That doesn't mean anything

This leads me to believe the author of this post doesn’t know what they’re talking about


So you routinely encounter photographs that have noticable areas where sensor did not receive any light during exposure? To the point where you feel not having completely unlit parts of a photo is a sign of over-exposure or filters? Are you an astrophotographer?


Right but the medium of a canvas isn't the same as an electronic screen.

What if you empower the user to control their device and use reasonable baselines that maximize legibility


16161d text on a background of fafafc is an 18:1 ratio, sufficient to meet WCAG AA criteria


Wow, I was searching for this exact link for more than a decade (!). I remember seeing it on HN when I was new here and couldn't find the article ever again. Thanks for sharing!


I'm wondering how many millions in developer productivity are being lost every minute.


It's not clear to me whether Anthropic's limitations are technical or merely contractual. Is Anthropic actually putting the limitations in their prompts, so that the model would refuse to answer a question on how to do certain things?

If so, that's a major problem. If the military is using it in some mission critical way, they can't be fighting the model to get something done. No such limitations would ever be acceptable.

If the limitations are contractual, then there is some room for negotiation.


> If the military is using it in some mission critical way, they can't be fighting the model to get something done. No such limitations would ever be acceptable.

You'd be surprised at what is considered acceptable. For example, being unable to repair your own equipment in battle is considered acceptable by decision-makers who accepted the restrictions.

https://www.warren.senate.gov/newsroom/press-releases/icymi-...


This looks great, but it's not clear to me how to use it for a practical task. I need to transcribe about 10 years worth of monthly meetings. These are government hearings with a variety of speakers. All the videos are on YouTube. What's the most practical and cost-effective way to get reasonably accurate transcripts?


If you use something like youtube-dlp you can download the audio from the meetings, and you could try things out in mistrals ai studio.

You could use their api (they have this snippet):

```curl -X POST "https://api.mistral.ai/v1/audio/transcriptions" \ -H "Authorization: Bearer $MISTRAL_API_KEY" \ -F model="voxtral-mini-latest" \ -F file=@"your-file.m4a" \ -F diarize=true \ -F timestamp_granularities="segment"```

In the api it took 18s to do a 20m audio file I had lying around where someone is reviewing a product.

There will, I'm sure, be ways of running this locally up and available soon (if they aren't in huggingface right now) but the API is $0.003/min. If it's something like 120 meetings (10 years of monthly ones) then it's roughly $20 if the meetings are 1hr each. Depending on whether they're 1 or 10 hours (or if they're weekly or monthly but 10 parallel sessions or something) then this might be a price you're willing to pay if you get the results back in an afternoon.

edit - their realtime model can be run with vllm, the batch model is not open


- get an API key for this service

- make sure you have a list of all these YouTube meeting URLs somewhere

- ask your preferred coding assistant to write you up a script that downloads the audio for these videos with yt-dlp & calls Mixtrals' API

- ????

- profit


> 10 years worth of monthly meetings

if it's 1 monthly video and thus 120 videos (or so) you could try recall (getrecall.ai not recall.ai that is a similar product with a similar name). They summarize youtube videos, but you get the transcript. AFAIK you cannot batch the processing and you have to add each video one by one, that's why 100 or 200 videos is doable but probably not thousands.


If they are on Youtube, try Gemini 3 Flash first. Use AI studio, it lets you insert YouTube videos into context.


Irked me too. I'm running for state representative in House District 9, which covers much of that area. The district is drop dead gorgeous and it's growing rapidly. There's a lot here.


Whoah, neat, congratulations or whatever you say to that. Good luck! Were you already involved in local politics in the area?


> whatever you say to that

Illegitimis non carborundum?


Thanks. That's exactly the right thing to say.


yes, quite a bit


Pretty much everyone on earth is proud of their home area. I've never really been to a place where people didn't describe their home as one of the most beautiful places on earth.

Earth has a lot of nice places. The flatness of Indiana can be calming and beautiful. But if you're driving on main roads, it's not very exciting, and it's not the kind of exciting people will fly across the world to see. And that was the point in the article.


Try going to west Oklahoma sometime. Not even the people who've never traveled think it's the most beautiful place on Earth.


I love my home town. I grew up there, and it was fine. I would never describe it as one of the most beautiful places on earth. I'm I proud of it? Only in the sense it's home and I have family there.


People do fly out to vacation in the part of Indiana he highlighted!

I'm not saying Indiana on the whole isn't flat. When I was growing up, their ad campaign was literally "there's more than corn in Indiana". I'm just saying, he couldn't have picked a worse map location to make the point with.


This area has a lot of hills.


Many areas have some minor hills, or some variant infinitely more interesting (hoosier myself, I regret moving here 10 years ago and may finally be bothered enough by it to deal with moving out in Spring). All in all, a state this big with less than 1,000 ft between the highest and lowest parts is just disappointingly flat, given the general lack of other natural attraction.

I'd say the caves in that region of the state are interesting, but options like Mammoth Cave are not that far away in KY (and they have better hills too).


178 billion? That's nothing. I did trillions just this morning. I went to the grocery store and picked an item off the shelf, effectively filtering out the trillions of other products that I could have picked but didn't.

They did not process 178 billion rows per second. They did a search that found something in a large data set by eliminating the parts of the data set that could not have contained the item. Same way I did by picking one grocery store and going straight to the shelf.


Hm, if I understand their product correctly they are building a DB and their filtering actually returns correct results.

So, the analogy doesn't really hold true unless you actually have these trillions of alternate products stored in your brain and manage to cite the matching subset on demand.


Seemingly, their way of thinking goes roughly like this:

If I have 10 billion rows in an SQL database, with a UNIQUE index, and do SELECT * FROM t1 WHERE pk=<number>, then I have “processed” 10 billion rows.

If I do 10k of these queries per second, I have processed 100T rows per second.


The performance hit you take for doing dynamic dispatch is real and measurable. It's a no-go if you're in a performance sensitive part of your app.


I write rust for no_std embedded devices and I just use stack_dst for this :)

Of course you‘re right if you can‘t alloc, you can‘t use Box or similar stuff.

Also you can avoid a lot of performance hits if you just implement most stuff on the dyn Trait directly!


Sure but most code paths are not hot. This is definitely premature optimization at the expense of everything else.


They do pay taxes. They each pay personal income tax on their $100k.


Still plenty left to pay their business taxes.


And if it’s employees? Do you ask them to contribute to the company’s taxes as well? After they’ve paid their own?


You pay them less money so you can afford your taxes, like literally everyone else.


... or you never have the startup, and just have 2 people unemployed and no produce nothing of use to anyone.

Which is what's happening.


If you can't afford your taxes your business model was flawed to begin with. In the above example there is more than enough money for it to still be worth doing.


You don't seem to understand the implications here. This requires bootstrapped startups to have gross margins substantially above incumbents in order to compete and not be cash flow negative after paying taxes.

It makes it substantially more cashflow intensive to build a new software business, which entrenches incumbents and reduces competition. It favors companies who have the cash to wait for the full 5-year depreciation cycle, i.e. the opposite of most bootstrapped startups.

Quick example:

$10,000 revenue

$8,000 paid to software developer

$1,000 paid to AWS

leaves $1000 in profit.

You received $10,000 into your business account, but spent 8000+1000 = $9000. Your business account has a balance of $1000 at the end of the year.

Section 174 means you can only deduct 1/10th of the $8000 in the first year, $800. Your total deductible business expenses for the year will be 800+1000 = $1800.

Your taxable profit for the year is 10000-1800 = $8,200. If your effective tax rate is 25% (generously low), you owe $2,050 in taxes.

You pay your $2,050 tax payment and your business account is overdrafted by $1,050. You need to add $1,050 from your personal funds to the business to cover the shortfall.

Your business was cash flow negative for the year. This makes it extremely difficult to bootstrap a software company.


Well, I said elsewhere, this effectively means (heavily) taxing anyone who's doing something new (meaning adding additional taxes on top of income tax). Essentially all of Europe does this, and people here often decry how they totally lack innovation across the entire continent.

I don't think these two are unrelated.

I also don't understand the objection. It's not like anyone's getting away from taxes due to this rule. This is about a temporary exemption from company income tax IF AND ONLY IF companies have someone pay income tax on that money (and only up to the point where that keeps makes sense). This "exemption" lets you not add 15%-20% tax on top of 40-55% income tax just to try a new business as a company.


So on the $200,000 it’s reasonable to you that they have to pay $120,000 ($80k income+$40k business) in taxes?


Two people earning $100k each would pay $28k income taxes each, totaling $56k. Where did this $80k come from?


What about state and city taxes? 80k might be a tad too high but in NYC on $100k, you would only take home around $65k.


The company doesn’t have any money to pay those taxes with.

If you give the company more money to use to cover its tax bill, then that further increases the company’s taxable income.


Pay less salary so you can pay your taxes? This isn't as complicated as y'all seem to want to make it.


That’s what makes them less competitive: they have to lower pay because they don’t have the cash on hand and revenue to amortize the deductions.


If you employ people on straight salaries, you are stuck. Not everyone is on share gravy train


Is there a paper or some other explanation of what they're doing under the hood?


Here is a quick overview, doesn't really explain the deep details though: https://www.youtube.com/watch?v=ikYsr6nvbdE

Basically think mixture of experts, but each expert is an encoder with it's own input tokenization/feature engineering part and then there is some machinery to parametrize these encoders based on natural language queries and stitch their outputs together to form unified vectors for objects in the index and the queries coming in.

Given that the whole framework is Apache 2.0, you can also check https://github.com/superlinked/superlinked and the docs: https://docs.superlinked.com/ https://docs.superlinked.com/concepts/overview

We have some proprietary tech besides that - mainly embedding models for things like location and then executors for the Superlinked DAG that help you run it on GPU-accelerated Spark and something analogous to that for streaming workloads - those proprietary things are how we make money, we call it Superlinked Cloud.


The issue in this case is being tested in other cases. It's about the "third party doctrine", the theory that the fourth amendment does not cover our information if it is in the possession of a third party.

https://nclalegal.org/press_release/ncla-asks-supreme-court-...

I blogged about this some time ago: https://ccleve.com/p/a-privacy-amendment


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: