My Android application have to show an image include 16 parts that download from server (for example the human
body), the user can touch some specific part of this image (for
example the head or the stomach) and this part have to be highlight in
some way (for example by displaying a dark - transparent layer around the image so
the highlight image can see clearly )
So how can I highlight an image when still loading the other images .
And can I support it from 2.1OS .
Thanks in advance !
Sorry for my poor english !
If your big image exists from smaller pics, Java has a class named MosaicDescriptor
The "Mosaic" operation creates a mosaic of two or more source images. This operation could be used for example to assemble a set of overlapping geospatially rectified images into a contiguous image. It could also be used to create a montage of photographs such as a panorama.
All source images are assumed to have been geometrically mapped into a common coordinate space. The origin (minX, minY) of each image is therefore taken to represent the location of the respective image in the common coordinate system of the source images. This coordinate space will also be that of the destination image.
Detail and reference
Related
I'm new to Android development. I have a library here that produces LCD images that come from a device through USB. These images are small, 120x80 pixel, and are 1-bit. I want to show these images on the UI.
All documents I've found explain how to show bitmaps from image files (PNG etc.) or how to show them from an app's resources. I did find out that I can add an ImageView to the UI. Then, for each incoming 120x80 pixel image, create a Bitmap instance, fill it with pixels, and assign it to the ImageView. However, I do not know if this yields the best performance here.
It is also important to keep in mind that nearest neighbor scaling must be used here. Bilinear filtering with such a small image produces a result that is too blurry.
From my experience with other languages, it seems wasteful to create a Bitmap instance for each incoming image. But perhaps I am wrong here.
Suggestions?
As i understand template matching finds areas of an image that match (are similar) to a template image (patch) according to
this
So i have two images of different dimension. And when i am trying to match my template image with the source image it usually marks the wrong area in the image.
Source Image
Template image
Output image
It works fine with threshold value if the images are of same dimension. What needs to be done to overcome this problem?? Or can i use template matching in this scenario??
Hi I am making a hangman, and I need to show image after image each time the player fails an attempt eg head first, then the body ect. I'm doing a layerdrawable but not how to scale it to the original image, how climb as I show the image and the images layer by layer to create the hangman?
I'd think the simplest way would be to decide what size your image is going to be. Then just use several PNG images with one body-part drawn on them and a clear background. As long as the images are all exactly over one another, you should see the familiar image of a stick figure drawn one part at a time.
Hope this helps.
I want to emboss a background image to the captured camera image.
i.e say i click a person's image in office, and pick the persons image and emboss it on the already existing hill background image,
So that the person's final image shows that he captured the image at hill side.
Can someone throw some idea on this or share code snippet to achieve this.
I didn't quite catch the embossing part even after reading the example. Image embossing is normally replacing each pixel of an image with a highlight or a shadow so the image area is raised to multiple levels having different depth.
But what you are asking is complex. If i am right, you want to merge the two images where the person from first image will be present in the hill side image.
The issue you will face here is object detection of a person in some random image. This is going to be really difficult. And moreover it will involve hardcore processing of the image with some math. At the end of the day, you do have to remember that what you are asking is for a mobile and not an desktop.
I suggest you to do some extensive research on object detection before proceeding with what you have in mind.
Here are some useful links
http://en.wikipedia.org/wiki/Outline_of_object_recognition
object detection android opencv
http://opencv.org/android
You can search SO for similar questions like this
In my application I use an image . I made some portion of the image transparent. Now How can I find the co-ordinates of the transparent portion of the image so that I am able to find the actual area of the non-transparent area.
I don't know the android SDK, but I doubt it has image manipulaiton built in. so using what ever language you are in, you need to open the image as data. Hopefully you can use an image library that will perform things like getting the size of the image. This will let you access the raw pixel data, which you can then check though till you find a pixel whos transparency is not equal to zero.
finding it in the rows is fairly easy, but the columns will take a bit more work. Remember that pixel data will be accessed as an array of length width * height in pixels.