Tag: Github

  • Apple releases Pico-Banana-400K dataset to AI research community

    Apple releases Pico-Banana-400K dataset to AI research community

    If nothing else, Apple is participating in the artificial intelligence community.

    On Wednesday, the company released Pico-Banana-400K, a highly curated 400,000-image research dataset which was built using Google’s Gemini-2.5 models.

    The research team, which published a study entitled “Pico-Banana-400K: A Large-Scale Dataset for Text-Guided Image Editing,” released the full 400,000-image dataset under a non-commercial research license. This allows anyone to use it and explore it, provided it’s for academic work or AI research purposes and not for commercial use.

    The work follows up on a Google project from a few months again, wherein Google released the Gemini-2.5-Flash-Image model, also known as Nanon-Banana, which is arguably the state-of-the-art when it comes to image editing models.

    Apple’s reseachers described the work as follows:

    “Despite these advances, open research remains limited by the lack of large-scale, high-quality, and fully shareable editing datasets. Existing datasets often rely on synthetic generations from proprietary models or limited human-curated subsets. Furthermore, these datasets frequently exhibit domain shifts, unbalanced edit type distributions, and inconsistent quality control, hindering the development of robust editing models.”

    From Apple’s end, the company pulled an unspecified number of real images from the OpenImages dataset, “selected to ensure coverage of humans, objects, and textual scenes.” The team then came up with a list of 35 different types of changes a user could ask the model to make, grouped into eight categories, which include the following:

    • Pixel & Photometric: Add film grain or vintage filter
    • Human-Centric: Funko-Pop–style toy figure of the person
    • Scene Composition & Multi-Subject: Change weather conditions (sunny/rainy/snowy)
    • Object-Level Semantic: Relocate an object (change its position/spatial relation)
    • Scale: Zoom in

    The researchers would then upload an image to Nano-Banana, alongside one of these prompts. Once Nano-Banana was done generating the edited image, the researchers would then have Gemini-2.5-Pro analyze the result, either approving it or rejecting it, based on instruction compliance and visual quality.

    The resulting object became Pico-Banana-400K, which “includes images produced through single-turn edits (a single prompt), multi-turn edit sequences (multiple iterative prompts), and preference pairs comparing successful and failed results (so models can also learn what undesirable outcomes look like).”

    The end result both acknowledges limitations in fine-grained spatial editing, layout extrapolation, and typography. Still, the researchers say that they hope Pico-Banana-400K will serve as “a robust foundation for training and benchmarking the next generation of text-guided image editing models.”

    The study can be found on arXiv, while the dataset itself is freely available for download and use on GitHub.

    Stay tuned for additional details as they become available.

    Via 9to5Mac, arXiv, and GitHub

  • Rumor: Apple looking to open aspects of Apple Intelligence up to third-party developers for iOS 19

    Rumor: Apple looking to open aspects of Apple Intelligence up to third-party developers for iOS 19

    Sometimes it’s the best idea to hand an idea over to third-party developers to get the most from it.

    According to Bloomberg, Apple will allow third-party developers to incorporate Apple Intelligence into their apps as of iOS 19. The company is said to be exploring new ways to do this, and while they won’t be able to access all of Apple’s AI model tech at first, initial efforts will consist of features and functions that use on-device processing, not elements that use cloud-based AI models.

    The sources add that the plan to offer AI model access to developers may be announced as part of WWDC 2025, on June 9. Such an effort could also kick the doors open towards integrating Apple Intelligence to a larger number of developers, and could perhaps create entirely new classes of apps that can’t be found anywhere else.

    While Apple certainly has a way to go with its consumer-focused AI efforts, it’s also done a fair amount on the developer side, too. In May, it was reported that Apple was working with Anthropic to integrate Claude Sonnet into an upgraded version of Xcode. It was being used internally to assist Apple engineers in writing, editing, and testing code. Apple has also been rumored to be prepping AI tools for third-party developers to use in Xcode as well as Swift.

    At present, Apple is dealing with some of its own internal woes about building Apple Intelligence, and internal complaints have mentioned hallucinations and slow performance, which hindered its release. When it comes to aiding developers, Apple is playing catch-up to others in the space. GitHub Copilot works with Xcode, as does ChatGPT.

    Stay tuned for additional details as they become available.

    Via AppleInsider and Bloomberg

  • Apple makes Swift Build available as open-source, posts developer files to GitHub

    Apple makes Swift Build available as open-source, posts developer files to GitHub

    If you’re a Swift Playgrounds fan as well as an open-source aficionado, you’re going to like this.

    Over the weekend, Apple announced that the company is making Swift Build open-source. Swift Build functions as the engine used by both Xcode as well as Apple’s internal projects.

    Apple offered the following comment on Saturday:

    “As Swift expands, there’s value in investing in matching cross-platform build tools that provide a powerful, consistent, and flexible experience across the ecosystem.

    As a foundational step in this new chapter of Swift build technologies, today Apple is open sourcing Swift Build, a powerful and extensible build engine that provides a set of build rules for building Swift projects. Swift Build is the engine used by Xcode, which supports millions of apps in the App Store as well as the internal build process for Apple’s own operating systems.”

    The company said that this would remove the confusion caused by having two different build packages, and pave the way for new features.

    For those interested in downloading and tinkering with it, Swift Build is now available on GitHub.

    Stay tuned for additional details as they become available.

    Via 9to5Mac, swift.org, and Daring Fireball

  • Code references hint at increased smart home device, Matter support in iOS 17.6 update

    Code references hint at increased smart home device, Matter support in iOS 17.6 update

    If you love controlling home accessories via your iPhone, you’re going to like Apple’s upcoming iOS 17.6 update.

    After examining a recent public code repository, the discovery of implementation source code for Matter, an Apple-backed smart home standard has been found. Matter is also open source, hence its public-facing code repository.

    The changes reference new features en route to unannounced versions of Apple’s platforms:

    Code additions reference support coming for these new features in the “middle of 2024” to iOS 17.6, macOS 14.6, watchOS 10.6, and tvOS 17.6 and contain the following:

    • release: “Middle of 2024”

    versions:
    iOS: “17.6”
    macOS: “14.6”
    watchOS: “10.6”
    tvOS: “17.6”

    It’s unknown as to exactly which smart home devices will be supported by the update, but the references seem to include more robust air quality monitoring, power management, and water valves. There appear to be references to CO, CO2, formaldehyde, PM10, PM2.5, PM1, and radon monitoring in the code strings.

    Other references point out devices such as robotic vacuum cleaners, which were part of the Matter 1.2 update.

    The commits on the Github repo were accredited to username bzbarsky-apple, or Boris Zbarsky. The Apple software references were only added in the last few days.

    Zbarsky, a former software engineer from Mozilla, has been with Apple for the past four years.

    Stay tuned for additional details as they become available.

    Via AppleInsider and Github (1, 2)

  • Rumor: Apple has begun developing AI tool to incorporate into Xcode

    Rumor: Apple has begun developing AI tool to incorporate into Xcode

    If you’ve been using AI to help with your coding tasks, you might like this.

    Apple is working on an updated version of Xcode that will include an AI feature to help generate code. According to Bloomberg, the AI tool will be similar to GitHub Copilot from Microsoft, which can generate code based on natural language requests and convert code from one programming language to another.

    Once implemented, the Xcode AI tool will be able to both predict and finish blocks of code, allowing developers to streamline their workflow. Apple is said to have begun testing the functionality internally with plans to release it to third-party software developers “as early as this year.”

    Apple is also testing AI-generated code for testing apps, and has asked some engineers to try these features out internally.

    This new feature will be added to other AI features that Apple plans to add to Siri and other apps. Some new features could include the option to generate playlists in Apple Music and create slideshows in Keynote, with Apple also working on improved Spotlight search capabilities. Search could encompass specific features in apps and might also provide responses to complex questions, with the feature built using large language models.

    Finally, Apple software chief Craig Federighi has asked employees to create as many new AI features as possible for the iOS 18, iPadOS 18, and macOS 15 operating systems.

    Stay tuned for additional details as they become available.

    Via MacRumors and Bloomberg