SmartEyeglass HeroSmartEyeglass Hero

透過式メガネ型端末 SmartEyeglass SED-E1

目の前の風景に情報が重ねて表示されるAR(拡張現実)の世界を体験しましょう。SmartEyeglass SED-E1*は、視線を大きく 変えることなくハンズフリーでさまざまな情報を取得できる 透過式メガネ型端末です。


ARグラス用ディプレイモジュール SED-100A はこちら


API overview

SmartEyeglass is a Sony accessory that extends the functionality of Android mobile devices.  The SmartEyeglass has a see-through binocular head-mount display, camera, sensors and a controller with a touch sensor and keys.

The Sony SmartExtension framework extends the Android SDK to work with Sony accessories. The SmartEyeglass SDK further extends the SmartExtension framework, defining a set of Java classes and data structures that enable your app to interact with and control the SmartEyeglass device. You use and extend the classes defined in the SmartEyeglassAPI library and in the sample projects in order to create an application for the SmartEyeglass device.

In order to get access to the SmartExtension and SmartEyeglass features, you must import the library projects into your own project. These projects are supplied with the SmartEyeglass SDK. The projects are provided as source code, rather than compiled JAR libraries. For more information, see the Software architecture page.

Features available through the SmartEyeglass API

You can use the SmartEyeglass API to show text and bitmaps on the device display, where it is overlaid on the user’s view of their surroundings. The UI that you define for your application consists of a top-level card that is the entry-point to your app, and a series of screens where you have full control over the display and can handle user input events.

Your app can also control the camera and microphone that are built in to the device, and receive state information from the hardware sensors on the device. You can, for example, get real-time information from the compass, gyroscope, and light sensor.

Your app defines handlers for events that can be received from the SmartEyeglass device.  You can define responses to user-input events, such as swipes, taps, and key-presses, and also for hardware events such as data being received from a sensor.

Your handlers will include logic for changing and adjusting what is displayed on the screen, and can also access the device’s sensors, and take actions based on things like the ambient brightness of the environment, and the user’s motions in the real world.

  • The sample projects provided with the SmartEyeglass SDK exercise and demonstrate the feature of the SmartEyeglass API; you can use these as a base for your own apps.
  • A set of guides provide usage information and recommendations for programming applications using the API and sample projects. See the Guides overview for more information.