Hacker Newsnew | past | comments | ask | show | jobs | submit | moochee's commentslogin

I’m curious how to auto-renew lets encrypt certs, do you mind sharing the tips? Thanks


I am using nginx-proxy (look it up on github) . It hasn't been that simple to put together, I will investigate MicroK8S in the future.


traefik (https://traefik.io/traefik) is also pretty good at this. I've used it to get certs auto-renewed for my projects.


This is super nice, thanks for sharing. Using gh issue => pr => deployment flow is good, but it would be awesome to have an optional local dev flow so the iterations can go even faster.


yes, now I found the time mainly caused by: 1. openAI API call, which could not be optimized for now 2. the time to build and deploy, which can be optimized by a preset setup


Is it similar to OpenAI GPT-4-Vision model?


What problem to solve with LoftQ? I can’t tell after looking at the readme.


from the paper: https://arxiv.org/pdf/2310.08659.pdf

"LoftQ aims to solve the problem of the discrepancy between the quantized and full-precision model in the context of quantization and LoRA fine-tuning for Large Language Models (LLMs). By simultaneously quantizing an LLM and finding a proper low-rank initialization for LoRA fine-tuning, LoftQ significantly enhances generalization in downstream tasks."

Bard: https://bard.google.com/chat/31e0a3bb74b29b3b

"Based on the abstract, LoftQ aims to solve the performance gap observed when applying both quantization and LoRA fine-tuning to a pre-trained Large Language Model (LLM).

Here's a breakdown of the problem and LoftQ's approach:

Problem:

Quantization: Reduces the precision of model weights to save memory and computation, but can lower accuracy. LoRA fine-tuning: Improves accuracy on specific tasks by adding a low-rank adapter, but can struggle with quantized models. Combined approach: Applying both quantization and LoRA fine-tuning often leads to a performance gap compared to full fine-tuning. LoftQ's solution:

Simultaneous quantization and LoRA initialization: LoftQ proposes a novel framework that quantizes the LLM while also finding a suitable low-rank initialization for LoRA. This helps bridge the gap between the quantized and full-precision model. Improved generalization: This approach improves the model's ability to generalize well on downstream tasks, especially in challenging memory-constrained settings. Evaluation and results:

LoftQ is tested on various NLP tasks like question answering and summarization. It outperforms existing quantization methods, particularly in low-precision scenarios like 2-bit and 2/4-bit mixed precision. Overall, LoftQ tackles the challenge of combining quantization and LoRA fine-tuning for LLMs, leading to better performance and efficiency, especially in resource-limited environments."


Thanks. LoftQ = Quantization + LoRA fine-tuning. What's the difference between LoftQ vs QLoRA then?


LoftQ = Quantization optimized for LoRA + Better LoRA Adaptor initialization + LoRA fine-tuning.


Nice work, thanks for making it! A few nice-to-haves: 1. it didn’t mention in readme that ollama has to be started manually in terminal. But I figured that out. 2. have a short video showing how it works in coding work, especially for people who never used gh copilot.


Yeah, this is my plans to add as well. I want to refine it a little bit from coding standpoint first.


Nice work. Tbh the tech stack should be open sourced so it could be easily picked up and improved by a larger community


Thanks. Can you be more specific?


More specific on what? Screen recording software and editing software?


I found its hard to mirror my iphone screen to my macbook, any idea?


probably an emulator is your best bet to record gameplay


Do you any example app in github?


You can use any app that supports OpenID Connect. Just set the client_id to the URL the app runs on.


Nice work. I saw that the json file will be published to CDN, the json file may contain some sensitive data e.g. company id. How do you handle that?


I’d like to recommend Insomnium, a 100% local-first Insomnia fork. https://archgpt.dev/insomnium


Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: