Hacker Newsnew | past | comments | ask | show | jobs | submit | more decodebytes's commentslogin

You can now generate reasoning with Tool calling or format pre-existing datasets to the OpenAI harmony format used for gpt-oss with DeepFabric:

https://github.com/lukehinds/deepfabric/discussions/334


I would love to learn more and have a try, I figure you can dump out to txt or csv -

you can raise and issue and I will certainly give it a go - or also reach me via the discord link on the main repo. Let's see what we can do.


sure, just starting to get some up on HF. A good example might be GSM8K as this shows the structured output where every result is strictly formatted - I am using this right now to train models and managaing to get a small qwen model up in the 60% range, which wildly is higher then llama2 and xAI Grok 1

GSM8K: https://huggingface.co/datasets/lukehinds/deepfabric-GSM8K-c...

also some others

infra failures reasoning / CoT: https://huggingface.co/datasets/lukehinds/deepfabric-devops-...

Medical (multi-turn): https://huggingface.co/datasets/lukehinds/deepfabric-7k-medi...

Programming challenges: https://huggingface.co/datasets/lukehinds/programming-challe...

If there is anything in particular you need, drop me a message or feel free to open an issue and I can create something for you.


Thanks, what LLMs were used to create these?


I think it was gpt4-mini, but local models do surprisingly well too.


Very good, and even better with the new DAG approach - we have been using great-expectations to bench and seeing very good diversity and low amounts of duplication - you check out one of the recent CoT examples here: https://huggingface.co/datasets/lukehinds/deepfabric-devops-...


This dataset disappeared. Did it move or get pulled for some reason? (glanced at it when you noted this and went back today to check it out and found a 404...)


Ah right, kiln - Deepfabric was originally named promptwright , and I can see kiln has copied over some of our code and used it for its synth-gen (which is a nice compliment!)

We are actually planning on moving to graphs now, which we are seeing better results with over trees, check it out if you also want to use them in kiln - but you might want to wait until we validate a little more and lift it out of experimental.

I think the key difference between the two since kiln adopted the same approach is the ability to generate reasoning / chain of thought and export to alpaca, chatml, etc - along with direct to unsloth.ai's formatting. I doubt we will have UI as its for running on backend systems and part of an ML pipeline along with being a library / SDK.


I personally wrote Kiln's SDG code myself -- no code was copied from here or anywhere else. Not sure where that claim is coming from, but it's not accurate.

I might have taken some of the prompts and modified them. I didn't recognize the new name, do recognize the old one.

Edit:

- just confirmed. No code copied. Prompts were originally from the Pluto library, then modified by the library above, then modified again by me for Kiln.

- And just to clarify, Kiln has had supported for chain of thought, reasoning, and all major export formats (ChatML/Unsloth/OpenAI/Hugging Face). Plus API integrations with Together, Fireworks, OpenAI, Google Vertex.

People should try both. I just want to clear on the origins of the code/prompts, and the feature set.


no worries, its not a big deal - I saw promptwrights name referenced in kilns source. Best of luck , looks like a cool project.


Line 1 makes it pretty clear:

    # The contents of this file are adapted from the promptwrite library (https://github.com/StacklokLabs/promptwright),
    # which was adapted from the pluto library (https://github.com/redotvideo/pluto).

https://github.com/Kiln-AI/Kiln/blob/d38a64b598bf21939263bed...

Curious how the OP "just confirmed. No code copied."


I read the code. I also remember writing the code and that comment.

As disclosed: some prompt strings were taken and modified, but none of the code was. The original strings are using a templating library that we don't support, so their code/strings wouldn't have worked in our codebase, nor would the wrapping code. Those interfaces/LOC are all unique. It's possible for some "content" to be taken (partial prompt strings), but zero code, and the statement "copied over some of our code and used it" to be incorrect.

Not trying to make a big deal of this, just clarifying these are separate libraries, with no shared code. Looks like the author saw the comment and assumed we used code (vs prompts); not a big deal, but not the case. Their work is super cool, and did inspire parts of my project.

Also worth noting, the library Pluto originated this prompt (as far as I know), and it's been tweaked/evolved many times over.


Hey There, this thread is getting derailed. Could you please create a separate post for your project and we let this one be for discussion of deepfabric, thanks!


Agreed, and sorry about that. Maybe edit the incorrect comment about "I can see kiln has copied over some of our code" for clarity. I get it was probably honest mistake, but hard not to reply when people are claiming I copied something I didn't. Great project, people go check out deepfabric!


Just searched for world record. 24 minutes!

Budimir Šobat (Croatia) at 56 years old no less!


Without breath up on supplemental oxygen, it is half that


A lot of this is plain naive and wrong.


How so?


Stacklok | Fully Remote | Multiple Open Roles

We are an innovative startup founded by the creators of Kubernetes, Sigstore and the folks who bootstrapped foundations such as the CNCF and OpenSSF.

Our mission is to revolutionize the software industry by providing a secure and trustworthy software supply chain. With our deep expertise in open-source technologies and commitment to enhancing software security, we are seeking a highly skilled and motivated individuals in multiple roles.

* Senior FrontEnd Engineer

* Senior Site Reliability Engineer

* Staff Product Manager

* Staff Security Software Engineer

* Staff Site Reliability Engineer

* Staff Software Engineer - Core Platforms and OSS

* Staff Software Engineer - ML/AI

Please apply at: https://www.stacklok.com/careers


Which LLM are you using?


Yep, currently gpt-3.5-16k or gpt-4. We wrote the example prompts in a relatively Llama-compatible way though (we actually started building this onto Llama 1 before switching to OpenAI as default), and make few assumptions about the LLM so it's easy to switch out. Mostly this is waiting behind us adding an option to pass in any LLM, and we're planning to add support for this.

Generally, we leave the LLM up to the user -- if OpenAI or Google is a no-go, then you probably are anyway in the territory of self-hosting or even self-training your LLM, which means you're fine setting up your own inference endpoints as well.


For the time being, it looks like you can choose between gpt-3.5-turbo-16k and gpt-4.

https://github.com/opencopilotdev/opencopilot/blob/d7aa8270d...


Looks like gpt4


> the richer a person is, the harder they are to corrupt

That is quite a lot view of non-rich citizens. A lot have morals as baseline for corruption, beyond the cash / assets we hold.


I view rich and non-rich people equally. Many are moral, many others are corruptible. Among the latter, it’s harder (more expensive) to corrupt a non-rich person


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: