I had gone through ImageTargets application, from which program i need to start developing my application, can u please help me
You need to connect your mobile to the PC just to install the application. After that, you can disconnect the mobile from the PC and run the application standalone on the device.
My question is
To track the live video into the program, we need to connect ur mobile to pc, is it right?
I don't understand the first question.
For the second, I can't tell you how long the battery will last for any given device. Running the camera preview will certainly have some small impact on battery life, but I don't believe we have measured this.
My question is
1) when i am tracking the video using the mobile phone connected to the pc
,long time should be tracked,how it can done? and
2)how much duration can a mobile phone will record the live video for htc desire?
Sorry, I don't really understand the question.
If you are asking which programs you should use for development, then the answer is this: Use Eclipse for your Java programming, and for building the application and running it on your Android device. For the native code you can use any IDE you would like (Eclipse, Visual Studio, etc.) but you build the code from the command line, using ndk-build. You can make changes to the native build in the Android.mk file, if you want to add additional classes, for instance.
I had gone through the examples specified in the qcar, i just confused how to start the application development from which program.
If you are not seeing any Build Targets on the New Android Project screen, there are two possible problems:
1) You have not installed any platforms for Android. In Eclipse, go to Window > Android SDK and AVD Manager. Choose Installed Packages, and make sure Platforms 2.1 and/or 2.2 are installed. If not, choose Available Packages and install them.
2) You have not pointed Eclipse to the correct Android location. In Eclipse, go to Window > Preferences. Choose Android, and set the SDK Location. Hit Apply at the bottom, and you should see a list of the Android platforms you have installed.
Finally, to run your application on a device, simply right-click on the project in the Package Explorer, and choose Run As > Android Application.
Compiliation is done i.e ndk-build is ok
1) i found apk in the bin folder, i can execute in my mobile?
2) i am unable to see the targets, in ths step of "create android apk"
Which version of the SDK are you using? We just released version 0.10.0 yesterday, before that the version number was 0.9.7. If you haven't grabbed the latest you might want to go ahead and do that, the Eclipse build is a little simpler.
When i am compiling the ImageTarget
error message is
cannot find -lshared
reply me asap
I think this should work, though I haven't tested that setup before. Just make sure that you point Eclipse to the correct Android location (Preferences > Android).
my ecllipse ide for c/c++ is in local disk (d:) drive,
remaining jdk,cygwin,andriodsdk,and qualcomm ar sdk are in local disk(c:) drive
is there any problem if the above softwares are like that?
You should be able to open it using Notepad.
How can i open the Teapot.h file? i tried using with wordpad, it is not opened
Sorry robertrose, I am starter to do coding in augmented reality, i know c++ & java well. i dont have any experience with opengl but i can understand.
Yes and yes.
But it sounds as though you really don't know what you're doing.
You're going to need experience with Java, C++, JNI, Android and openGL ES to create your application.
This stuff is really not as easy as clicking and dragging things together.
Just take some time and go read the "Developer Guide" and the "Get Started" document.
Ok, my question is
1) Is it possible to add my own virtual content, such as googles by using qcar sdk?
2)I can use htc desire? for developing the above application
I should warn you that this stuff isn't easy. You really need to work through the sample applications and understand the code there before you can continue. You're going to need a basic understanding of OpenGL ES to render any 3D objects.
First, you will have to convert your model to a format that you can easily read in and feed to OpenGL. Look at the Teapot.h file included in the Shared sample project. This is one possible format for storing your model data.
Some users in the forum here have been using this script to convert a .obj model file to a header file like Teapot.h:
Once you have your model in header form, I suggest trying to replace the Teapot model in the VirtualButtons project. You may as well work with that project, since you eventually want virtual button support in your application. Once you get one custom model rendering correctly, we can talk about how to switch models on a button press.
1)I am asking that i need to change the virtual content i.e googles from time to time. is it possible?
2)How can i store the virtual content of various no of googles?
3) I dont want to change the color of the virtual object but it should change the new virtual content wat i had stored
The QCAR SDK uses Natural Feature Tracking to track image targets. You preprocess the image using our online Target Management System (https://ar.qualcomm.com/qdevnet/projects). Using this tool, you can see the set of features the tracker will be looking for. At runtime, the tracker compares the camera feed with this known set of features, and tries to find the target in the image. Once it locks on the target, the tracker can figure out where the target is in relation to the device. This is given to the developer as a Pose matrix, and can be used to place 3D content on the target.
Can i know which
1) image recognization technique is used to recognize the Image Target.
2)what was the technique used for knowing the position and orientation of the external point reference or Image Target.
Sure, that shouldn't be too difficult. Take a look at the VirtualButtons sample project, and VirtualButtons.cpp in particular. In that sample, when the buttons are tapped the model's texture is changed. For your application, you will also want to change the model that is drawn when the button is tapped. The basic code structure is similar though, that project should be a good starting point.
My question is
1) I need to place various virtual objects such as googles one by one, so that i
need to create forward and backward virtual buttons.
if forward button is clicked, the next virtual object should be displayed
if backward button is clicked, previous virtual object should be come
is it possible to write the code in c++ to get above functionality
No, in fact the JNI isn't something separate that your download. It is simply the interface you use for communicating between Java and native code.
You can write all of your rendering code in native (C/C++) and bypass the need to communicate with Java. Simply make changes to the ImageTarget.cpp file to try it out.
I need to place various virtual objects such as googles in my application, so
1) Is i need to download java native interface(JNI) to get above functionality?
Start with the ImageTargets sample project. There are two files of interest:
ImageTargets.cpp contains all the OpenGL ES rendering code. Teapot.h contains the geometry to be rendered. You can replace the arrays in Teapot.h with your own model data.
Some users have had luck with the following script, for converting a .obj model to the format in the header file:
You should be able to save your model to a .obj file using any standard modeling tool.
i need to create virtual objects such as googles with different colors, so that i need to place them on the human.
so my question is
1)how the googles are created using opengl libraries, can u have any example.
1) Yes, you can track custom image targets. Use the My Trackables tool on this site to create your custom target.
2) Yes, you can place virtual objects on the target in the live video. This runs on your handheld Android device.
3) Yes, you do need OpenGL libraries for rendering 3D virtual content. These are included in the Android NDK.
I suggest trying one of the sample projects included with the SDK, it will give you a better understand of how all this works.
By using qcar sdk, is it possible to do
1) custom markers tracking in the live video
2) placing virtual objects on the custom marker patterns in live video
3) for placing virtual objects in live video, we need opengl libraries, all these are included in qcar sdk ?
That just depends on whether or not you want to use Unity. There are two pathways for using the QCAR SDK:
1) Download the qcar-sdk and follow the directions in the Get Started guide. With this method you write native code in C++, use ndk-build to build a native shared library, and use Eclipse to build an Android application including that shared library.
- or -
The documentation for using the Unity extension is under the AR Extension for Unity 3 topic on the home page:
I have a Image with me in A4 paper, i need to track that image in live video , is i need to downloand the Ar extension for unity3 ? for tracking the image which i had? plz relpy me
Yes, the SDK works with Windows 7.
Hello, is qualcomm ar sdk will run on windows7 ?. plz reply me
I just ask that you read through the documentation once before asking questions, most of your answers are there :)
Cygwin is a unix-like environment for Windows, for running command-line tools. To build applications with the QCAR SDK, you must use the ndk-build command that comes with the Android NDK. This must be run from a command line, and in Windows you use Cygwin. In Mac or Linux you can use the basic command line tool that comes with the operating system.
wat is the use of cygwin, can u tell me
sorry kim, i am very new to this, so i had many questions to ask. please dont mind.
You use the online My Trackables tool for uploading images to create targets:
Please understand, the QCAR SDK is for writing MOBILE applications. You cannot use it to write desktop applications. You cannot test the applications on your desktop either, you can only test on the mobile device. The entire application experience is handheld.
You write your code on your personal computer, and cross-compile for the Android OS, using Eclipse and tools provided by Google. The end result is an application that only runs on the Android OS, on an Android device.
how can we create custom markers of my own desire and how that code will be passed to the program to track that marker in the live video? plz reply to my questions it will help me
While writing the program in ecllipse with c++ language, we will track the live video using your phone attached to the usb is it? And how can we see the live video in the personal computer? please let me know and it will be help me to do the applications
The SDK is for writing the programs which are then installed onto the phone. The SDK itself does not go on the phone.
The whole AR experience is on the phone, not the PC.
But kim, the sdk is in our pc, how can we track the live video from the personal computer without having camera attached to it.
The entire AR application runs on the mobile phone. It renders live video in the background (from the device's camera), performs tracking, and renders your content on the targets. The entire AR experience takes place on the device.
Thanks kim, mobile phone attached to the usb is used for taking the live video is it? reply me
You must install the QCAR SDK on your personal computer, it can be downloaded from here:
Please follow the instructions in the Get Started guide, on the left-hand side of that page. You will be building applications in Eclipse and deploying them to your Android device.
I am asking that, to develop Ar applications we need to install qcar sdk in the mobile phone or personal computer ? plz let me know
Applications built with the QCAR SDK only run on Android devices. Please see the Device Compatibility sticky for a list of supported devices:
Hello siva, i need to install qcar sdk in the andorid mobile phone or in the personal computer? to develop ar applications
Are you sure you want to delete this message?
Are you sure you want to delete this conversation?
To reset your password please enter the email address associated with your account. An email will be sent to you with instructions on how to complete changing your password.