2 research outputs found

    On-device LLM-driven Application Function Calling for Natural Language Interaction

    Get PDF
    Large language models can enable application programmers to translate user instructions in natural language into a sequence of corresponding function calls; however, currently LLMs that provide such capabilities are housed in the cloud. Use of remotely hosted LLMs can introduce operational latency, affect reliability, and can be infeasible in certain cases. This disclosure describes the use of an on-device LLM to enable app developers to support natural language interactions with users without having to implement code within their applications. Per the techniques, with user permission, a user command in natural language as received by an application is provided to the on-device LLM that generates an appropriate sequence of function calls that can be executed one at a time, with each function receiving the output of the prior function, to perform the requested task. The techniques enable a deep level of natural language conversational interaction within any application by utilizing application and user context as inputs to the LLM that generates the function calls. The use of application-defined function definitions makes it easier for developers to control the actions of the model, providing greater stability than other approaches such as direct UI control

    JaxPruner: A concise library for sparsity research

    Full text link
    This paper introduces JaxPruner, an open-source JAX-based pruning and sparse training library for machine learning research. JaxPruner aims to accelerate research on sparse neural networks by providing concise implementations of popular pruning and sparse training algorithms with minimal memory and latency overhead. Algorithms implemented in JaxPruner use a common API and work seamlessly with the popular optimization library Optax, which, in turn, enables easy integration with existing JAX based libraries. We demonstrate this ease of integration by providing examples in four different codebases: Scenic, t5x, Dopamine and FedJAX and provide baseline experiments on popular benchmarks.Comment: Jaxpruner is hosted at http://github.com/google-research/jaxprune
    corecore