Android :: How To Use Map To Zoom With Hand Gesture
Apr 15, 2009How do you use the Android API to zoom in and out maps with the 2 finger gestures like iPhone?
View 2 RepliesHow do you use the Android API to zoom in and out maps with the 2 finger gestures like iPhone?
View 2 RepliesI've got GoClever brand tablet (model orion 70L). Everything works fine but when I want to zoom image/pdf using two finger gesture it is very slow. I neet do repeat it ten times to zoom image as I want to. Is there any option/application to make it "faster"?
P.S If I open google maps (preinstalled in tablet) this gesture works fine as in for example Angry Birds. So I think that its a software problem...
I have just updated my hero phone and I was surprised it worked ok, normally things for me go wrong. However before when using the camera the zoom bar appeared on the left hand edge of the camera display and I could zoom using the trackball.
Now it has an icon on the left to go to my albums and no zoom scale. I can select it from the menu but again I cant zoom using the trackball. Anyone know how to get this back if possible?
Is Zooming is possible to do in android for Image view.I have a images in full screen view, i need to zoom in and zoom out for that images present in image view.Can any body give me suggestions regarding this.
View 19 Replies View RelatedI want to know how we can zoom a image that is in ImageView.
View 3 Replies View RelatedI want to run a dj tango app on a hand-held device. It'll need to run Python (obviously) and will write its data to an SQLite database. Are there any tablets available that will let me do this? Specifically, if I bought an Android tablet, would I have to/be able to install Linux instead, or would I be able to run it under Android?
View 2 Replies View RelatedI would like to know what is meant by gestures in typical mobiles more specifically android ones. Android supports gesture recognition.
Is a gesture termed as when user holds the device and shakes it(say upwards, downwards or side- side)?
Is a gesture termed as when a finger is placed on the screen and traced UP,DOWN,RIGHT,LEFT? If so what is the difference between touch screen and gestures.
I am confused between 1) and 2) option.
What is a gesture exactly?
My employer is having me work on an android application with very specific UI requirements. Specifically, we need an interface that looks exactly like a TabWidget interface. However, having the tabs on the left hand side is an absolute must. We aren't going to be deploying this application to any android phones, its for an in house device, and as such don't have any problem violating any design considerations that the android platform might have.
We already have a working prototype that hacks together the functionality we need using a combination of focus listeners and list views. It isn't pretty, and we aren't very confident in it just yet, but it works for the time being.
What we'd really like is to throw all of our prototype UI code away in exchange for using the built in TabWidget. However, since TabWidget is hardcoded to only work with the tabs on top, this isn't really an option.
So, we are hoping that someone out there has a patch, or set of patches, or perhaps a custom class, that handles the TabWidget functionality with the tabs on the side instead?
The relevant code for the TabWidget is here: http://www.google.com/codesearch/p?hl=en#uX1GffpyOZk/core/java/android/widget/TabWidget.java&q=android%20package:git://android.git.kernel.org%20lang:java%20tabwidget&sa=N&cd=1&ct=rc
And the specific function that is hardcoded for being tabs on top is
CODE:.......................
I create some gestures and they are working, but if I add a button the gesture doesn't work over the button. I split the display resolution to make different gestures and i get x, y coordinates. If i make a gesture starting with the button the coordinates are both 0, only under them i get normal coordinates. It's like the coordinates are starting only after the button.
View 1 Replies View RelatedI have to add a gesture to my listview,i want to implement the same functionality of contact application.
when i left swipe it should send a message,right swipe it should call. can anyone help me how to do those gesture detection... i have implemented it in various other views... but i couldn't do for listView... i don't what going worng...
My code is:
CODE:...........
I want to know that android has a touchscreen requirement about hardware and software.And how does touch gesture processing where, for example, in driver layer or framework or application,
View 4 Replies View RelatedI got a crash with the following stack trace:
CODE:.................
Line 130 in my code is: ...........
What's wrong and what checks can I perform on gesture before sending it to recognize to avoid this crash?
I had developed a simple gesture recognition library for Android (not for accelerometer; I mean touchscreen - finger gestures). I had to abandon it for quite a while, but recently I was able to clean it a bit and port it to 1.0 SDK. It is released under GPL. Here are some sample videos:
http://www.youtube.com/watch?v=C7A6OHWfSOE [url]
http://android-wydiwyg.blogspot.com/
Me and some of my friends have made a gesture recognition application for android. It allows you to customize gestures that will call, text, or email one of your contacts or launch an application. Please check out our blog and demonstration video and comment to let us know what you think.
I got pretty good at using the Palm shorthand for entering text. I'm pretty sure that a similar "shorthand keyboard" would work out well on the Android phones, but I haven't seen any apps that will replace the keyboard with a "shorthand" pad.
View 2 Replies View RelatedI have hooked in a ScaleGestureDetector to an OnTouchListener as instructed in Android documentation. For some reason the ScaleGestureDetector does not always detect end of a pinch gesture. This happens mostly when pinching fast from large to small.
The effect is that after I have released both fingers the detector does not fire the gesture end event. If I touch the screen with one afterwards it still thinks I'm continuing the scale gesture (keeps firing onScale events). I have to do another gesture to get the detector to fire end event.
I have added logs to the OnTouchListener and when the scale gesture gets stuck I still get motion events normally when using one finger and event.getPointerCount() is 1.
I have made sure that I don't have any other listeners intervening. View that has the onTouchListener is not the only view on screen but this effect happens also when I'm very careful to begin and end the gesture inside this one view.
Is there any way to improve the end detection?
Or if there a way for me to manually set the ScaleGestureDetector to fire onScaleEnd and change scaleGestureDetector.isInProgress() to false?
What I want is to have an options menu where the user can choose to navigate the menu between:
1) touching a button and then pressing down on the trackball to select it, and
2) drawing predefined gestures from Gestures Builder
As it stands now, I have created my buttons with OnClickListener and the gestures with GestureOverlayView. Then I select starting a new Activity depending on whether the using pressed a button or executed a gesture. However, when I attempt to draw a gesture, it is not picked up. Only pressing the buttons is recognized. The following is my code:
CODE:..............
I am using gesture detector to catch "flings" and using a view flipper to change the screen when this happens. Some of my child views contain list views. The the gesture detector wont recognize a swipe if you swipe on the list view. But it will recognize it if it is onTop of TextView's or ImageView's. Is there a way to implement it so that it will recognize the swipes even if they are on top of another view that has a ClickListener?
View 3 Replies View RelatedWhat Android Api is used to get the scrolling to the left or to the right on the start screen on Android?
View 4 Replies View RelatedI am adding gesture recognition to my app. I have added the view as described in the Android Developers Gestures article but when it comes to adding:
CODEE:..........
where do I put this in my code, do I have to create a new class for it, or can I have it in an inner class, or does it not need a class at all?! I have a set up similar to Lunar Lander which comprises of two files, one of which is a thread that handles pretty much all the physics and drawing of the game. The other file begins the thread and saveInstanceState method. Furthermore, what type is mLibrary?! I cannot find out anywhere! I imagine I will put the OnGesturePerformed method in my thread as this is where I handle all keyUp and Down events.
I have been having some 4G issues so I was taking a look at my signal strength today and noticed that if I cover the top/back of the phone (i.e. over the camera) with my hand the signal drops from about ~90 dBm to ~100 dBm. As soon as I move my hand down goes back to about 90 dBm. FYI, this is happening while I am on 3G not 4G. Are others EVO owners seeing the same drop in signal due to hand placement?
View 15 Replies View RelatedI can't seen to find the neck or hand strap holes on this phone which is common in a lot of phones and mp3 players. Are they on somewhere on the phone?
View 1 Replies View RelatedI have the same problem with Gesture Builder on the emulator and when I build the app myself and install om my G1.
It does not save the gestures correctly on my sdcard. It does not create a folder called "gestures" but instead creates a file called "gestures" with a size on about 2kb. It does not grow when I create more gestures. Is this really correct, because I thought a folder would be created in which each gesture would be stored in different files.
I look in code for CreateGestureActivity.java and here is where the magic of saving happens:
final GestureLibrary store = GestureBuilderActivity.getStore(); store.addGesture(name.toString(), mGesture); store.save();
I want to implement the same behavior of the notification view in the official Facebook app.
The "notifications" tab are at the bottom and can drag/drop via fingers to full screen. How can I do that? I've tried it via ViewFlipper and Animation. But no success. The app "Zedge" has the same in the "search" function. Via drag/drop you can open the "search" view.
I want to implement the same behaviour of the notification view in the official Facebook app.
The "notifications" tab are at the bottom and can drag/drop via fingers to full screen.
How can i do that?
I've tried it via ViewFlipper and Animation.... But no success.
Does anyone know how we can do this?
The app "Zedge" has the same in the "search" function. Via drag/drop you can open the "search" view.
Edit: Here's the original thread by igl007, w/ video:
http://androidforums.com/support-troubleshooting-incredible/70366-unresponsive-screen-when-car-dock.html[/QUOTE]
I am a brand new developer to Android. I wrote a pretty simple application that uses the gesture library. The application itself seems to run fine in the simulator. However, when I installed the application on my Google Dev Phone 1 (G1) last night, it doesn't seem to work. I've upgraded my G1 with the 1.6 image from the htc developer site. I see the first screen/activity for about a second and the application exits without and errors. Here is what my manifest looks like:
<manifest xmlns:android="http://schemas.android.com/apk/res/android"
package="com.blahblah" android:debuggable="true"
android:versionCode="1"
android:versionName="1.0">
<uses-sdk android:minSdkVersion="4" />
<uses-permission android:name="android.permission.WRITE_EXTERNAL_STORAGE" />
<uses-permission android:name="android.permission.READ_CONTACTS">
</ uses-permission>
<uses-permission android:name="android.permission.CALL_PHONE">
</uses- permission>
<application android:icon="@drawable/icon" android:label="@string/ application_name"> <activity android:name="Gesture1Activity"
android:label="@string/application_name">
<intent-filter>
<action android:name="android.intent.action.MAIN" />
<category android:name="android.intent.category.LAUNCHER" />
</intent-filter>
</activity>
<activity android:name="Gesture2Activity"
android:label="@string/label_gesture_list" />
<activity android:name="Gesture3Activity"
android:label="@string/label_create_gesture" />
</application>
</manifest>
Here's the output for logcat:
D/AndroidRuntime(860): D/AndroidRuntime(860): >>>>>>>>>>>>>> AndroidRuntime START <<<<<<<<<<<<<< D/AndroidRuntime(860): CheckJNI is OFF D/AndroidRuntime(860): --- registering native functions --- I/jdwp (860): received file descriptor 20 from ADB D/ddm-heap(860): Got feature list request I/ActivityManager(101): Starting activity: Intent { flg=0x10000000 cmp=com.blahblah/.Gesture1Activity } D/AndroidRuntime(860): Shutting down VM D/dalvikvm(860): DestroyJavaVM waiting for non-daemon threads to exit E/AndroidRuntime(860): ERROR: thread attach failed D/dalvikvm(860): DestroyJavaVM shutting VM down D/dalvikvm(860): HeapWorker thread shutting down D/dalvikvm(860): HeapWorker thread has shut down D/jdwp (860): JDWP shutting down net... D/jdwp (860): +++ peer disconnected I/dalvikvm(860): Debugger has detached; object registry had 1 entries D/dalvikvm(860): VM cleaning up D/dalvikvm(860): LinearAlloc 0x0 used 677572 of 4194304 (16%) W/InputManagerService(101): Window already focused, ignoring focus gain of: com.android.internal.view.IInputMethodClient$Stub $Proxy@439a9eb8
My View subclass is can be written one of two ways, I've tried both. One implements OnGestureListener, the other doesn't. In both cases, the Activity registers the view for context menus when the view is created and registers itself, the activity, as the context menu listener.In the case where OnGestureListener is *not* implemented, context menus work. onCreateContextMenu() and onContextItemSelected() are both called and the menu works as expected. In the case where OnGestureListener *is* implemented by the view subclass, the context menu processing functions in the activity are never called. Problem is, I need both. Without OnGestureListener, I don't get fling events. With it, I don't get context menus. need both.
View 6 Replies View RelatedThe App I'm trying to work on requires that when a special combination of keys are pressed ( irrespective of the keypad being locked ) a service / app is triggered.
How can I achieve this? My app's whole idea hinges on this particular part to work.
I am not here to take a "no" for an answer. If you are going to tell me that its not possible, atleast include an explanation as to why its not possible.
If (in Android) you wanted to write an alarm app where the user entered in the time for the alarm by swiping the hours & minutes up and down, what's the easiest way of achieving this? Is it the Swipe or Fling command?
Are there any good tutorials out there for this?