Skip to content

importRyan/Colors-We-See

Repository files navigation

Colors We See

Simulates color blindness on your camera feed using Metal filters.

The primary objective is to explore the latest Composable Architecture beta and UGC-driven App Clip cards ahead of a larger App Clip-focused project.

Test Flight Structure Screenshots To Dos

Structure

Quick overview of reducer/SPM module composition of this few day old hobby project:

[[iOS/App Clip Targets]]
  (Thin clients instantiating AppRootScene)

[[Features Package]]
  [Root] 
  └── AppRoot
     └── Bootstrap
  Integration tests cover deferred deeplinking and sad bootstrap paths

  [Tabs]
  Composes Camera, Learn, and Photos Tabs

  [CameraTab]
  └── Camera
     ├── CameraPermissions
     └── AppClipFullAppDownload
  Integration tests hit async permissions flow, app clip upgrade prompt

  [Common]
  Analytics, feature flags, and other shared specifics

[[Clients Package]]
  └── Analytics
  └── App Clip
  └── Feature Flags
  └── Launch Environment
  └── Vision Simulation (Camera + Metal Compute)

[[ColorVision Package]]
  (Reusable GPU and CPU color blindness simulation tools)

Screenshots

ColorsWeSee

Tasks

  • Setup XCC (pre-build script to populate Firebase secrets, test plans)
  • Remote config (version check)
  • Symbol stripping
  • App icon
  • QC TestFlight
  • Submit barebones 1.0
  • User selected image simulation + export (PhotosTab)
  • UGC sharing (snap photos, user selected images)
  • Slicker CameraTab
  • UGC-driven App Clip card
  • Non-fatals
  • App Clip code generation and export
  • macOS target
  • macOS ScreenCaptureKit tab
  • Shortcuts
  • Learn tab
  • Improve barebones video interface / exposure / focus
  • Metal simulation cross fade/rolling fade
  • Animate camera permissions screen
  • visionOS target (limited to LearnTab, PhotosTab)
  • String Catalog accessor generation