• 0 Posts
  • 33 Comments
Joined 1 year ago
cake
Cake day: September 25th, 2023

help-circle
  • Super, thanks again for taking the time to do so.

    I can’t remember if I shared this earlier but I’m jolting down notes on the topic in https://fabien.benetou.fr/Content/SelfHostingArtificialIntelligence so I do also invest time on the topic. Yet my results have also been… subpar so I’m asking as precisely as I can how others actually benefit from it. I’m tired of seeing posts with grand claims that, unlike you, only talk about the happy path in usage. Still, I’m digging not due to skepticism as much as trying to see what can actually be leveraged, not to say salvaged. So yes, genuine feedback like yours is quite precious.,

    I do seem to hear from you and others that to kickstart what would be a blank project and get going it can help. Also that for whatever is very recurrent AND popular, like common structures, it can help.

    My situation though is in prototyping where documentation is sparse, if even existent, and working examples are very rare. So far it’s been a bust quite often.

    Out of curiosity, which AI tools specifically do you use and do you pay for them?

    PS: you mention documentation is both cases, so I imagine it’s useful when it’s very structured and when the user can intuit most of how something works, closer to a clearly named API with arguments than explaining the architecture of the project.


  • Thanks for that, was quite interesting and I agree that completion too early (even… in general) can be distracting.

    I did mean about AI though, how you manage to integrate it in your workflow to “automate the boring parts” as I’m curious which parts are “boring” for you and which tools you actual use, and how, to solve the problem. How in particular you are able to estimate if it can be automated with AI, how long it might take, how often you are correct about that bet, how you store and possibly share past attempts to automate, etc.





  • FWIW I did try a lot (LLMs, code, generative AI for images, 3D models) in a lot of ways (CLI, Web based, chat bot) both locally and using APIs.

    I don’t use any on a daily basis. I find it exciting that we can theoretically do a lot “more” automatically but… so far the results have not been worth the efforts. Sadly some of the best use cases are exactly what you highlighted, i.e low effort engagement for spam. Overall I find that either working with a professional (script writer, 3D modeler, dev, designer, etc) is a lot more rewarding but also more efficient which itself makes it cheaper.

    For use cases where customization helps while quality does matter much due to scale, i.e spam, then LLMs and related tools are amazing.

    PS: I’d love to hear the opinion of a spammer actually, maybe they also think it’s not that efficient either.







  • Right, and I mentioned CUDA earlier as one of the reason of their success, so it’s definitely something important. Clients might be interested in e.g Google TPU, startups like Etched, Tenstorrent, Groq, Cerebras Systems or heck even design their own but are probably limited by their current stack relying on CUDA. I imagine though that if backlog do keep on existing there will be abstraction libraries, at least for the most popular ones e.g TensorFlow, JAX or PyTorch, simply because the cost of waiting is too high.

    Anyway what I meant isn’t about hardware or software but rather ROI, namely when Goldman Sachs and others issue analyst report saying that the promise itself isn’t up to par with actual usage for paying customers.



  • I’m not sure if you played PCVR in the Summer but imagine that in a tiny room… it’s just way too hot. Again I’m NOT saying it’s good, or bad, I’m only saying you made assumption about OP usage. I’m not sure if you tried CloudXR but basically, it works and it’s not that complex to setup (e.g 1h) so it’s relatively faster and cheaper than building and owning a gaming PC.

    I don’t understand why you are even arguing about a legitimate usage.





  • Stuff like LLMs or ConvNets (and the likes) can already be used to do some pretty amazing stuff that we could not do a decade ago, there is really no need to shit rainbows and puke glitter all over it.

    I’m shitting rainbows and puking glitter on a daily basis BUT it’s not against AI as a field, it’s not against AI research, rather it’s against :

    • catastrophism and fear, even eschatology, used as a marketing tactic
    • open systems and research that become close
    • trying to lock a market with legislation
    • people who use a model, especially a model they don’t even have e.g using a proprietary API, and claim they are an AI startup
    • C-levels decision that anything now must include AI
    • claims that this or that skill is soon to be replaced by AI with actually no proof of it
    • meaningless test results with grand claim like “passing the bar exam” used as marketing tactics
    • claims that it scales, it “just needs more data”, not for .1% improvement but for radical change, e.g emergent learning
    • for-profit (different from public research) scrapping datasets without paying back anything to actual creators
    • ignoring or lying about non renewable resource consumption for both training and inference
    • relying on “free” or loss leader strategies to dominate a market
    • promoting to be doing the work for the good of humanity then signing exclusive partnership with a corporation already fined for monopoly practices

    I’m sure I’m forgetting a few but basically none of those criticism are technical. None of those criticism is about the current progress made. Rather, they are about business practices.


  • Their valuation is because there’s STILL a lineup a mile long for their flagship GPUs.

    Genuinely curious, how do you know where the valuation, any valuation, come from?

    This is an interesting story, and it might be factually true, but as far as I know unless someone has actually asked the biggest investor WHY they did bet on a stock, nobody why a valuation is what it is. We might have guesses, and they might even be correct, but they also change.

    I mentioned it few times here before but my bet is yes, what you did mention BUT also because the same investors do not know where else do put their money yet and thus simply can’t jump boats. They are stuck there and it might again be become they initially though the demand was high with nobody else could fulfill it, but I believe that’s not correct anymore.


  • Unfortunately it’s part of the marketing, thanks OpenAI for that “Oh no… we can’t share GPT2, too dangerous” then… here it is. Definitely interesting then but now World shattering. Same for GPT3 … but through exclusive partnership with Microsoft, all closed, rinse and repeat for GPT4. It’s a scare tactic to lock what was initially open, both directly and closing the door behind them through regulation, at least trying to.