Hacker Newsnew | past | comments | ask | show | jobs | submit | peadarohaodha's commentslogin

Would love to have a good selection of keyboard shortcuts! Power Vibe Kanban


Good point - there are a few random ones already but I will ask a coding agent to implement this comprehensively


We've found you can get an order of magnitude improvement in the amount of labelled data you need - but there is some variance based on the difficulty of the problem. Because you are retraining the model in tandem with the data labelling process, there is additional compute associated to an active learning powered labelling process versus just selecting the data at random to label next. But this additional compute consideration is almost always outweighed by the saving of human time spent labelling.


I have a question on the compute aspect regarding your business model, hope I’m not being to nosy..

I tried HL, the experience was stellar (well done!) and it made me think...

To get AL working with a great user experience you need quite a bit of compute. How are you thinking about your margins, e.g the cost to produce what you’re offering versus what customers will pay for it?


Thanks for the feedback! It's a good question re compute. There are some fun engineering and ML research challenges that we are constantly iterating on that are related to this. A few examples - how to most efficiently share compute resources in a JIT manner (e.g. GPU memory) during model serving for both training and inference (where the use case and privacy requirements permit) - how to construct model training algorithms that operate in a more online manner effectively (so you don't have to retrain on the whole dataset when you see new examples) - how to significantly reduce the model footprint (in terms of memory and flops) of modern deep transformer models given they are highly over-parameterised and can contain a lot of redundancy.

this stuff helps us a lot on the margins point!


adding to what Raza said - a consideration for both active learning and weak supervision is the need to construct a gold standard labelled dataset for model validation and testing purposes. At Humanloop, in addition to collecting training data, we are also using a form of active learning to speed up the collection of an unbiased test data set.

Another consideration on the weak supervision side (for the snorkel style approach of labelling functions) is that creating labelling functions can be a relatively technical task, which may not be well suited for non-technical domain expert annotators for providing feedback to the model


Adding to what Raza as said - to your point on "real active learning" hardly working. I would be interested to hear what approaches you took? We've found that the quality of the uncertainty estimate for your model is quite important for active learning to work well in practise. So applying good approximations for the model uncertainty for modern sized transformer models (like BERT) is an important consideration


Irrc we were using a Bayesian classification model on top of of fixed pretrained features from transfer, something along the lines of refitting a GP every time the number of classes changed. This was images as opposed to text, and after an epoch classification was ~ok but during training (eg the active bit) we didn't see much benefit.


Our default deployment option is cloud first for both training and inference at the moment, but we have thought about the ability for users to export a trained model. Either exporting the model parameters in some standardised format, or a compiled predict function, or a docker image that encapsulates a full inference service, etc. So if you could use this kind of export within your application, this would allow on-premise inference. This is something we could probably make available pretty quickly if necessary for your use case.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: