I did nothing and I’m all out of ideas!

  • 0 Posts
  • 16 Comments
Joined 2 years ago
cake
Cake day: June 11th, 2023

help-circle
  • I’ve never used oobabooga but if you use llama.cpp directly you can specify the number of layers that you want to run on the GPU with the -ngl flag, followed by the number.

    So, as an example, a command (on linux) from the directory you have the binary, to run its server would look something like: ./llama-server -m "/path/to/model.gguf" -ngl 10

    Another important flag that could interest you is -c for the context size.

    This will put 10 layers of the model on the GPU, the rest will be on RAM for the CPU.

    I would be surprised if you can’t just connect to the llama.cpp server or just set text-generation-webui to do the same with some setting.

    At worst you can consider using ollama, which is a llama.cpp wrapper.

    But probably you would want to invest the time to understand how to use llama.cpp directly and put a UI in front of it, Sillytavern is a good one for many usecases, OpenWebUI can be another but - in my experience - it tends to have more half baked features and the development jumps around a lot.

    As a more general answer, no, the safetensor format doesn’t directly support quantization, as far as I know


  • AFAIK it is still a tuning of llama 3[.1], the new Base models will come with the release of 4 and the “Training Data” section of both the model cards is basically a copy paste.

    Honestly I didn’t even consider the fact they would not be giving Base models anymore before reading this post and, even now, I don’t think this is the case. I went to search the announcements posts to see if there was something that could make me think about it being a possibility, but nothing came out.

    It is true that they released Base models with 3.2, but there they had added a new projection layer on top of that, so the starting point was actually different. And 3.1 did supersede 3…

    So I went and checked the 3.3 hardware section and compare it with the 3 one, the 3.1 one and the 3.2 one.

    3 3.1 3.2 3.3
    7.7M GPU hours 39.3M GPU hours 2.02M GPU hours 39.3M GPU hours

    So yeah, I’m pretty sure the base of 3.3 is just 3.1 and they just renamed the model in the card and added the functional differences. The instruct and base versions of the models have the same numbers in the HW section, I’ll link them at the end just because.

    All these words to say: I’ve no real proof, but I will be quite surprised if they will not release the Base version of 4.

    Mark Zuckerberg on threads

    Link to post on threads
    zuck a day ago
    Last big AI update of the year:
    •⁠ ⁠Meta AI now has nearly 600M monthly actives
    •⁠ ⁠Releasing Llama 3.3 70B text model that performs similarly to our 405B
    •⁠ ⁠Building 2GW+ data center to train future Llama models
    Next stop: Llama 4. Let’s go! 🚀

    Meta for Developers

    Link to post on facebook
    Today we’re releasing Llama 3.3 70B which delivers similar performance to Llama 3.1 405B allowing developers to achieve greater quality and performance on text-based applications at a lower price point.
    Download from Meta: –

    Small note: I did delete my previous post because I had messed up the links, so I had to recheck them, whoops






  • disable this system security feature temporarily,

    This should be - if I’m not mistaken - possible using the pip env var I posted about earlier, like this:

    PIP_BREAK_SYSTEM_PACKAGES=1 sudo apt install howdy

    Or exporting it for the current shell, before running the installation

    export PIP_BREAK_SYSTEM_PACKAGES=1

    But I personally highly discourage it, because - AFAIK - if it even works it will mess up the deps in your system.


  • I’m no python expert but reading around it seems your only real solution is using a virtual environment, through pipx or venv as you already had found out, or using the

    --break-system-packages
    
    * Allow pip to modify an EXTERNALLY-MANAGED Python installation
    
      (environment variable: `PIP_BREAK_SYSTEM_PACKAGES`)
    

    pip flag which, as the name suggest, should be avoided.

    EDIT: After rereading I got your problem better and I was trying to read the source for Howdy to see how to do it, so far no luck.



  • Mechanize@feddit.ittoLinux@lemmy.mlproton VPn
    link
    fedilink
    arrow-up
    0
    ·
    edit-2
    1 year ago

    Considering you are not using the Flatpak anymore it is, indeed, strange. The only reasons I can think of are: your network manager is using the wrong network interface to route your traffic ( if you go on an ip checking site like for example ipinfo do you see yours or the VPN’s IP?) or that you have WebRTC enabled and the broadcaster is getting your real ip through that.

    For the first case it can get pretty complicated, but it is probably an error during the installation of the VPN app or you set up multiple network managers and it gets confused on which one to configure. You should also enable the Advanced Kill Switch in the configuration.

    For the second case you could try adding something like the Disable WebRTC add-on for firefox and check if it works. Remember to enable it for Private Windows too.

    The last thing I can think of is that you allowed the broadcaster to get your real geolocation (in firefox it should be a small icon on the left of the address bar), or you are leaking some kind of information somewhere: there are a bunch of site that check for ip leak, but I don’t know if that goes too deep for you.
    If you want to check anyway the first two results from DDG are browserleaks and ipleak. Mullvad offered one too but it is currently down.

    EDIT: If you enable the Advanced Kill Switch, and the app is working correctly, internet will not work while you are not connected to a VPN server or until you disable the switch again, so pay attention to that.







  • Because, as pointed in the page, Servo is being developed as a(n embeddable) Rendering Engine, not as a full blown end user Browser.
    Its alternatives are not Chrome, Safari or Firefox, but Webkit, Blink and Gecko

    There’s an example GUI called Servoshell, but it is more of a testing ground and example on how to embed the engine in an app than a serious alternative to anything currently in the market.

    Already this kind of work is difficult and daunting. Adding to it a full GUI would make it completely impossible for the current size and financial backing Servo has.

    Big words aside it just means that Servo wants to be only one of the parts that compose a real browser: the one that takes HTML, Javascript, WASM and translates them into the things you see on your monitor. All the user facing functionality are left to the devs of the app that embed it.