- iOS/Objective-C 元类和类别
- objective-c - -1001 错误,当 NSURLSession 通过 httpproxy 和/etc/hosts
- java - 使用网络类获取 url 地址
- ios - 推送通知中不播放声音
我有一个在启动时使用文字转语音的应用程序。一切似乎都运行良好,运行该应用程序时我没有遇到任何问题。但是,每次应用程序启动时,我都会收到一个 LogCat 错误,提示我的文本到语音转换中存在泄漏的 ServiceConnection。虽然这目前不会导致功能问题,但我担心如果我继续忽略该错误,将来会发生什么。我怀疑这只是错位的编码,但我没有确定错位的经验。我已经包含了 LogCat 和有问题的 java。
07-31 10:16:50.812: E/ActivityThread(27785): Activity com.example.com.proto1.menu has leaked ServiceConnection android.speech.tts.TextToSpeech$1@40d68398 that was originally bound here
07-31 10:16:50.812: E/ActivityThread(27785): android.app.ServiceConnectionLeaked: Activity com.example.com.proto1.menu has leaked ServiceConnection android.speech.tts.TextToSpeech$1@40d68398 that was originally bound here
07-31 10:16:50.812: E/ActivityThread(27785): at android.app.LoadedApk$ServiceDispatcher.<init>(LoadedApk.java:932)
07-31 10:16:50.812: E/ActivityThread(27785): at android.app.LoadedApk.getServiceDispatcher(LoadedApk.java:827)
07-31 10:16:50.812: E/ActivityThread(27785): at android.app.ContextImpl.bindService(ContextImpl.java:1109)
07-31 10:16:50.812: E/ActivityThread(27785): at android.content.ContextWrapper.bindService(ContextWrapper.java:370)
07-31 10:16:50.812: E/ActivityThread(27785): at android.speech.tts.TextToSpeech.initTts(TextToSpeech.java:517)
07-31 10:16:50.812: E/ActivityThread(27785): at android.speech.tts.TextToSpeech.<init>(TextToSpeech.java:483)
07-31 10:16:50.812: E/ActivityThread(27785): at com.example.com.proto1.menu.onActivityResult(menu.java:255)
07-31 10:16:50.812: E/ActivityThread(27785): at android.app.Activity.dispatchActivityResult(Activity.java:4581)
07-31 10:16:50.812: E/ActivityThread(27785): at android.app.ActivityThread.deliverResults(ActivityThread.java:2817)
07-31 10:16:50.812: E/ActivityThread(27785): at android.app.ActivityThread.handleSendResult(ActivityThread.java:2864)
07-31 10:16:50.812: E/ActivityThread(27785): at android.app.ActivityThread.access$1000(ActivityThread.java:122)
07-31 10:16:50.812: E/ActivityThread(27785): at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1057)
07-31 10:16:50.812: E/ActivityThread(27785): at android.os.Handler.dispatchMessage(Handler.java:99)
07-31 10:16:50.812: E/ActivityThread(27785): at android.os.Looper.loop(Looper.java:132)
07-31 10:16:50.812: E/ActivityThread(27785): at android.app.ActivityThread.main(ActivityThread.java:4126)
07-31 10:16:50.812: E/ActivityThread(27785): at java.lang.reflect.Method.invokeNative(Native Method)
07-31 10:16:50.812: E/ActivityThread(27785): at java.lang.reflect.Method.invoke(Method.java:491)
07-31 10:16:50.812: E/ActivityThread(27785): at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:844)
07-31 10:16:50.812: E/ActivityThread(27785): at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:602)
07-31 10:16:50.812: E/ActivityThread(27785): at dalvik.system.NativeStart.main(Native Method)
import java.util.Locale;
import android.os.Bundle;
import android.app.Activity;
import android.content.Intent;
import android.view.Menu;
import android.view.MenuItem;
import android.view.View;
import android.view.View.OnClickListener;
import android.widget.Toast;
import android.speech.tts.TextToSpeech;
import android.speech.tts.TextToSpeech.OnInitListener;
import android.support.v4.app.NavUtils;
@SuppressWarnings("unused")
public class mainj extends Activity implements OnInitListener {
private TextToSpeech myTTS;
// status check code
private int MY_DATA_CHECK_CODE = 0;
// setup TTS
public void onInit(int initStatus) {
// check for successful instantiation
if (initStatus == TextToSpeech.SUCCESS) {
if (myTTS.isLanguageAvailable(Locale.US) == TextToSpeech.LANG_AVAILABLE)
myTTS.setLanguage(Locale.US);
} else if (initStatus == TextToSpeech.ERROR) {
Toast.makeText(this, "Sorry! Text To Speech failed...",
Toast.LENGTH_LONG).show();
}
}
@Override
public void onCreate(Bundle savedInstanceState) {
super.onCreate(savedInstanceState);
setContentView(R.layout.loadscreen);
Intent checkTTSIntent = new Intent();
checkTTSIntent.setAction(TextToSpeech.Engine.ACTION_CHECK_TTS_DATA);
startActivityForResult(checkTTSIntent, MY_DATA_CHECK_CODE);
Thread logoTimer = new Thread() {
public void run() {
try {
try {
sleep(3000);
speakWords("main menu loaded");
} catch (InterruptedException e) {
// TODO Auto-generated catch block
e.printStackTrace();
}
Intent menuIntent = new Intent("android.intent.action.MENU");
startActivity(menuIntent);
}
finally {
finish();
}
}
};
logoTimer.start();
}
// speak the user text
private void speakWords(String speech) {
// speak straight away
if (myTTS != null) {
myTTS.speak(speech, TextToSpeech.QUEUE_FLUSH, null);
}
}
// act on result of TTS data check
protected void onActivityResult(int requestCode, int resultCode, Intent data) {
if (requestCode == MY_DATA_CHECK_CODE) {
if (resultCode == TextToSpeech.Engine.CHECK_VOICE_DATA_PASS) {
// the user has the necessary data - create the TTS
myTTS = new TextToSpeech(this, this);
} else {
// no data - install it now
Intent installTTSIntent = new Intent();
installTTSIntent
.setAction(TextToSpeech.Engine.ACTION_INSTALL_TTS_DATA);
startActivity(installTTSIntent);
}
}
}
}
package com.example.com.proto1;
import android.app.Activity;
import android.content.Intent;
import android.media.MediaPlayer;
import android.os.Bundle;
import android.view.View;
import android.widget.Button;
import android.content.pm.PackageManager;
import android.content.pm.ResolveInfo;
import android.speech.RecognizerIntent;
import android.view.View.OnClickListener;
import android.widget.ArrayAdapter;
import android.widget.EditText;
import android.widget.ListView;
import android.widget.Toast;
import android.speech.tts.TextToSpeech;
import java.util.ArrayList;
import java.util.List;
import java.util.Locale;
@SuppressWarnings("unused")
public class menu extends Activity implements TextToSpeech.OnInitListener,
OnClickListener {
// defined
TextToSpeech mTts;
public static final int VOICE_RECOGNITION_REQUEST_CODE = 1234;
// remember to include a listview on the xml or the voice recognition code
// will not work
public ListView mList;
// TTS object
Button speakButton, infoButton, voiceButton, talkButton;
// TTS object
public TextToSpeech myTTS;
// status check code
public int MY_DATA_CHECK_CODE = 0;
// setup TTS
public void onInit(int initStatus) {
// check for successful instantiation
if (initStatus == TextToSpeech.SUCCESS) {
if (myTTS.isLanguageAvailable(Locale.US) == TextToSpeech.LANG_AVAILABLE)
myTTS.setLanguage(Locale.US);
} else if (initStatus == TextToSpeech.ERROR) {
Toast.makeText(this, "Sorry! Text To Speech failed...",
Toast.LENGTH_LONG).show();
}
}
@Override
protected void onCreate(Bundle aboutmenu) {
super.onCreate(aboutmenu);
setContentView(R.layout.mainx);
SpeakingAndroid speak = new SpeakingAndroid();
VoiceRecognition voiceinput = new VoiceRecognition();
// get a reference to the button element listed in the XML layout
speakButton = (Button) findViewById(R.id.btn_speak);
infoButton = (Button) findViewById(R.id.aboutbutton);
voiceButton = (Button) findViewById(R.id.voicebutton);
talkButton = (Button) findViewById(R.id.talk);
// listen for clicks
infoButton.setOnClickListener(this);
speakButton.setOnClickListener(this);
talkButton.setOnClickListener(this);
// check for TTS data
Intent checkTTSIntent = new Intent();
checkTTSIntent.setAction(TextToSpeech.Engine.ACTION_CHECK_TTS_DATA);
startActivityForResult(checkTTSIntent, MY_DATA_CHECK_CODE);
// calling method
voiceinputbuttons();
// Check to see if a recognition activity is present
// if running on AVD virtual device it will give this message. The mic
// required only works on an actual android device//
PackageManager pm = getPackageManager();
List<ResolveInfo> activities = pm.queryIntentActivities(new Intent(
RecognizerIntent.ACTION_RECOGNIZE_SPEECH), 0);
if (activities.size() != 0) {
voiceButton.setOnClickListener(this);
} else {
voiceButton.setEnabled(false);
voiceButton.setText("Recognizer not present");
}
}
public void quitApp() {
speakWords("closing app");
finish();
}
public void recogMenu() {
speakWords("voice recognition menu");
startActivity(new Intent("android.intent.action.RECOGNITIONMENU"));
}
public void informationMenu() {
speakWords("information screen");
startActivity(new Intent("android.intent.action.INFOSCREEN"));
}
public void mainMenu() {
speakWords("main menu");
startActivity(new Intent("android.intent.action.MENU"));
}
// creating method
public void voiceinputbuttons() {
speakButton = (Button) findViewById(R.id.btn_speak);
mList = (ListView) findViewById(R.id.list);
}
// respond to button clicks
public void onClick(View v) {
switch (v.getId()) {
// use switch case so each button does a different thing
// accurately(similar to an if statement)
case R.id.btn_speak:
String words1 = speakButton.getText().toString();
// speakwords(xxxx); is the piece of code that actually calls the
// text to speech
speakWords(words1);
Intent voiceIntent = new Intent(
"android.intent.action.RECOGNITIONMENU");
startActivity(voiceIntent);
break;
case R.id.aboutbutton:
String words2 = infoButton.getText().toString();
speakWords(words2);
Intent infoIntent = new Intent("android.intent.action.INFOSCREEN");
startActivity(infoIntent);
break;
case R.id.voicebutton:
speakWords("Speak Now");
startVoiceRecognitionActivity(); // call for voice recognition
// activity
break;
case R.id.talk:
speakWords("This is the main menu.");
break;
}
}
// speak the user text
// setting up the speakWords code
public void speakWords(String speech) {
// speak straight away
myTTS.speak(speech, TextToSpeech.QUEUE_FLUSH, null);
}
/**
* Fire an intent to start the speech recognition activity.
*/
public void startVoiceRecognitionActivity() {
Intent intent = new Intent(RecognizerIntent.ACTION_RECOGNIZE_SPEECH);
intent.putExtra(RecognizerIntent.EXTRA_LANGUAGE_MODEL,
RecognizerIntent.LANGUAGE_MODEL_FREE_FORM);
intent.putExtra(RecognizerIntent.EXTRA_PROMPT,
"Speech recognition demo");
startActivityForResult(intent, VOICE_RECOGNITION_REQUEST_CODE);
}
/**
* Handle the results from the recognition activity.
*/
@Override
public void onActivityResult(int requestCode, int resultCode, Intent data) {
if (requestCode == VOICE_RECOGNITION_REQUEST_CODE
&& resultCode == RESULT_OK) {
// Fill the list view with the strings the recognizer thought it
// could have heard
ArrayList<String> matches = data
.getStringArrayListExtra(RecognizerIntent.EXTRA_RESULTS);
mList.setAdapter(new ArrayAdapter<String>(this,
android.R.layout.simple_list_item_1, matches));
// matches is the result of voice input. It is a list of what the
// user possibly said.
// Using an if statement for the keyword you want to use allows the
// use of any activity if keywords match
// it is possible to set up multiple keywords to use the same
// activity so more than one word will allow the user
// to use the activity (makes it so the user doesn't have to
// memorize words from a list)
// to use an activity from the voice input information simply use
// the following format;
// if (matches.contains("keyword here") { startActivity(new
// Intent("name.of.manifest.ACTIVITY")
if (matches.contains("information")) {
informationMenu();
}
if (matches.contains("info screen")) {
informationMenu();
}
if (matches.contains("info")) {
informationMenu();
}
if (matches.contains("about")) {
informationMenu();
}
if (matches.contains("home")) {
mainMenu();
}
if (matches.contains("menu")) {
mainMenu();
}
if (matches.contains("home screen")) {
mainMenu();
}
if (matches.contains("speak")) {
startActivity(new Intent("android.intent.action.SPEAK"));
}
if (matches.contains("close")) {
quitApp();
}
if (matches.contains("stop")) {
quitApp();
}
if (matches.contains("finish")) {
quitApp();
}
if (matches.contains("voice")) {
recogMenu();
}
if (matches.contains("recognition")) {
recogMenu();
}
if (matches.contains("voice recognition")) {
recogMenu();
}
}
// still in the onActivityResult: This is for the text to speech part
if (requestCode == MY_DATA_CHECK_CODE) {
if (resultCode == TextToSpeech.Engine.CHECK_VOICE_DATA_PASS) {
// the user has the necessary data - create the TTS
myTTS = new TextToSpeech(this, this);
} else {
// no data - install it now
Intent installTTSIntent = new Intent();
installTTSIntent
.setAction(TextToSpeech.Engine.ACTION_INSTALL_TTS_DATA);
startActivity(installTTSIntent);
}
}
super.onActivityResult(requestCode, resultCode, data);
}
}
08-01 08:19:18.942: E/ActivityThread(7310): Activity com.example.com.proto1.menu has leaked ServiceConnection android.speech.tts.TextToSpeech$1@40d68418 that was originally bound here
08-01 08:19:18.942: E/ActivityThread(7310): android.app.ServiceConnectionLeaked: Activity com.example.com.proto1.menu has leaked ServiceConnection android.speech.tts.TextToSpeech$1@40d68418 that was originally bound here
08-01 08:19:18.942: E/ActivityThread(7310): at android.app.LoadedApk$ServiceDispatcher.<init>(LoadedApk.java:932)
08-01 08:19:18.942: E/ActivityThread(7310): at android.app.LoadedApk.getServiceDispatcher(LoadedApk.java:827)
08-01 08:19:18.942: E/ActivityThread(7310): at android.app.ContextImpl.bindService(ContextImpl.java:1109)
08-01 08:19:18.942: E/ActivityThread(7310): at android.content.ContextWrapper.bindService(ContextWrapper.java:370)
08-01 08:19:18.942: E/ActivityThread(7310): at android.speech.tts.TextToSpeech.initTts(TextToSpeech.java:517)
08-01 08:19:18.942: E/ActivityThread(7310): at android.speech.tts.TextToSpeech.<init>(TextToSpeech.java:483)
08-01 08:19:18.942: E/ActivityThread(7310): at com.example.com.proto1.menu.onActivityResult(menu.java:255)
08-01 08:19:18.942: E/ActivityThread(7310): at android.app.Activity.dispatchActivityResult(Activity.java:4581)
08-01 08:19:18.942: E/ActivityThread(7310): at android.app.ActivityThread.deliverResults(ActivityThread.java:2817)
08-01 08:19:18.942: E/ActivityThread(7310): at android.app.ActivityThread.handleSendResult(ActivityThread.java:2864)
08-01 08:19:18.942: E/ActivityThread(7310): at android.app.ActivityThread.access$1000(ActivityThread.java:122)
08-01 08:19:18.942: E/ActivityThread(7310): at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1057)
08-01 08:19:18.942: E/ActivityThread(7310): at android.os.Handler.dispatchMessage(Handler.java:99)
08-01 08:19:18.942: E/ActivityThread(7310): at android.os.Looper.loop(Looper.java:132)
08-01 08:19:18.942: E/ActivityThread(7310): at android.app.ActivityThread.main(ActivityThread.java:4126)
08-01 08:19:18.942: E/ActivityThread(7310): at java.lang.reflect.Method.invokeNative(Native Method)
08-01 08:19:18.942: E/ActivityThread(7310): at java.lang.reflect.Method.invoke(Method.java:491)
08-01 08:19:18.942: E/ActivityThread(7310): at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:844)
08-01 08:19:18.942: E/ActivityThread(7310): at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:602)
08-01 08:19:18.942: E/ActivityThread(7310): at dalvik.system.NativeStart.main(Native Method)
import java.util.ArrayList;
import java.util.List;
import java.util.Locale;
import android.app.Activity;
import android.content.Intent;
import android.content.pm.PackageManager;
import android.content.pm.ResolveInfo;
import android.os.Bundle;
import android.speech.RecognizerIntent;
import android.speech.tts.TextToSpeech;
import android.speech.tts.TextToSpeech.OnInitListener;
import android.view.View;
import android.view.View.OnClickListener;
import android.widget.ArrayAdapter;
import android.widget.Button;
import android.widget.ListView;
import android.widget.Toast;
public class Infoactive extends Activity implements OnClickListener,
OnInitListener {
public static final int VOICE_RECOGNITION_REQUEST_CODE = 1234;
public ListView mList;
public Button speakButton, about, voicetest;
// TTS object
public TextToSpeech myTTS;
// status check code
public int MY_DATA_CHECK_CODE = 0;
// setup TTS
public void onInit(int initStatus) {
// check for successful instantiation
if (initStatus == TextToSpeech.SUCCESS) {
if (myTTS.isLanguageAvailable(Locale.US) == TextToSpeech.LANG_AVAILABLE)
myTTS.setLanguage(Locale.US);
} else if (initStatus == TextToSpeech.ERROR) {
Toast.makeText(this, "Sorry! Text To Speech failed...",
Toast.LENGTH_LONG).show();
}
}
@Override
protected void onCreate(Bundle savedInstanceState) {
super.onCreate(savedInstanceState);
setContentView(R.layout.infoscreen);
// get a reference to the button element listed in the XML layout
about = (Button) findViewById(R.id.button1);
voicetest = (Button) findViewById(R.id.button2);
// listen for clicks
about.setOnClickListener(this);
// check for TTS data
Intent checkTTSIntent = new Intent();
checkTTSIntent.setAction(TextToSpeech.Engine.ACTION_CHECK_TTS_DATA);
startActivityForResult(checkTTSIntent, MY_DATA_CHECK_CODE);
voiceinputbuttons();
// Check to see if a recognition activity is present
PackageManager pm = getPackageManager();
List<ResolveInfo> activities = pm.queryIntentActivities(new Intent(
RecognizerIntent.ACTION_RECOGNIZE_SPEECH), 0);
if (activities.size() != 0) {
voicetest.setOnClickListener(this);
} else {
voicetest.setEnabled(false);
voicetest.setText("Recognizer not present");
}
}
// respond to button clicks
@SuppressWarnings("unused")
public void onClick(View v) {
switch (v.getId()) {
case R.id.button1:
// get the text entered
Button buttonText = (Button) findViewById(R.id.button1);
String words = buttonText.getText().toString();
speakWords("EyePhone will help vizoolly impaired users to avoid obstacles, identify faces, and recognize objects.");
break;
case R.id.button2:
speakWords("Speak Now");
startVoiceRecognitionActivity();
break;
}
}
// speak the user text
public void speakWords(String speech) {
// speak straight away
myTTS.speak(speech, TextToSpeech.QUEUE_FLUSH, null);
}
public void voiceinputbuttons() {
speakButton = (Button) findViewById(R.id.btn_speak);
mList = (ListView) findViewById(R.id.list);
}
/**
* Fire an intent to start the speech recognition activity.
*/
public void startVoiceRecognitionActivity() {
Intent intent = new Intent(RecognizerIntent.ACTION_RECOGNIZE_SPEECH);
intent.putExtra(RecognizerIntent.EXTRA_LANGUAGE_MODEL,
RecognizerIntent.LANGUAGE_MODEL_FREE_FORM);
intent.putExtra(RecognizerIntent.EXTRA_PROMPT,
"Speech recognition demo");
startActivityForResult(intent, VOICE_RECOGNITION_REQUEST_CODE);
}
/**
* Handle the results from the recognition activity.
*/
@Override
public void onActivityResult(int requestCode, int resultCode, Intent data) {
if (requestCode == VOICE_RECOGNITION_REQUEST_CODE
&& resultCode == RESULT_OK) {
// Fill the list view with the strings the recognizer thought it
// could have heard
ArrayList<String> matches = data
.getStringArrayListExtra(RecognizerIntent.EXTRA_RESULTS);
mList.setAdapter(new ArrayAdapter<String>(this,
android.R.layout.simple_list_item_1, matches));
// matches is the result of voice input. It is a list of what the
// user possibly said.
// Using an if statement for the keyword you want to use allows the
// use of any activity if keywords match
// it is possible to set up multiple keywords to use the same
// activity so more than one word will allow the user
// to use the activity (makes it so the user doesn't have to
// memorize words from a list)
// to use an activity from the voice input information simply use
// the following format;
// if (matches.contains("keyword here") { startActivity(new
// Intent("name.of.manifest.ACTIVITY")
if (matches.contains("information")) {
startActivity(new Intent("android.intent.action.INFOSCREEN"));
}
if (matches.contains("info screen")) {
startActivity(new Intent("android.intent.action.INFOSCREEN"));
}
if (matches.contains("info")) {
startActivity(new Intent("android.intent.action.INFOSCREEN"));
}
if (matches.contains("about")) {
startActivity(new Intent("android.intent.action.INFOSCREEN"));
}
if (matches.contains("home")) {
startActivity(new Intent("android.intent.action.MENU"));
}
if (matches.contains("menu")) {
startActivity(new Intent("android.intent.action.MENU"));
}
if (matches.contains("home screen")) {
startActivity(new Intent("android.intent.action.MENU"));
}
if (matches.contains("speak")) {
startActivity(new Intent("android.intent.action.SPEAK"));
}
if (matches.contains("close")) {
finish();
}
if (matches.contains("stop")) {
finish();
}
if (matches.contains("finish")) {
finish();
}
if (matches.contains("voice")) {
Intent voiceIntent = new Intent(
"android.intent.action.RECOGNITIONMENU");
startActivity(voiceIntent);
}
if (matches.contains("recognition")) {
Intent voiceIntent = new Intent(
"android.intent.action.RECOGNITIONMENU");
startActivity(voiceIntent);
}
if (matches.contains("voice recognition")) {
Intent voiceIntent = new Intent(
"android.intent.action.RECOGNITIONMENU");
startActivity(voiceIntent);
}
}
if (requestCode == MY_DATA_CHECK_CODE) {
if (resultCode == TextToSpeech.Engine.CHECK_VOICE_DATA_PASS) {
// the user has the necessary data - create the TTS
myTTS = new TextToSpeech(this, this);
} else {
// no data - install it now
Intent installTTSIntent = new Intent();
installTTSIntent
.setAction(TextToSpeech.Engine.ACTION_INSTALL_TTS_DATA);
startActivity(installTTSIntent);
}
}
super.onActivityResult(requestCode, resultCode, data);
}
}
最佳答案
您应该在完成后调用 myTTS.shutdown()
,例如在您的 onDestroy
中。
关于java - 泄漏的 ServiceConnection Android 文本到语音,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/11742350/
无法清除输出流:java.net.SocketException:关闭失败:EBADF(错误文件编号) 有没有人得到这个,有没有解决这个异常的方法? 最佳答案 这可能意味着您没有写入该文件的权限或者它
我很困惑。 我有一个我认为需要实现服务的应用程序。该服务特定于应用程序;当应用程序实际死亡或被手动终止时,该服务就会消失。应用程序将与服务保持持续通信,最好是通过服务本身的 Activity 调用方法
调用时出现以下异常: requestWindowFeature(Window.FEATURE_ACTION_BAR_OVERLAY); 2089-2089/? E/ActivityThread﹕ Se
这是调用我的服务的类: public class TicketList extends ListActivity { private ArrayList alTickets = new ArrayLi
我尝试将代码集成到 Blundell's in-app purchasing example 中进入我自己的应用程序,它几乎可以正常工作。我做了一个名为 shop.java 的 Activity ,它
我在我的应用程序中使用 TTS。在 onPuase 方法期间,我使用以下代码停止引擎。问题是每当我暂停 Activity 说使用主页按钮时,我都会收到粘贴在日志猫部分中的错误。我从日志 cat 中了解
我在搞乱 Android 服务,我发现当我绑定(bind)到服务时,ServiceConnection.onServiceConnected() 会被相当可预测地调用。 但是,我的 onService
我有一个带有私有(private)字符串属性的MainActivity。在我的代码中,我有以下内容: private ServiceConnection mTransactionServiceConn
我打开一个应用程序做了一些事情,然后应用程序崩溃了。我再次打开它并收到此错误。我想知道这个错误的原因和解决方法。谁能帮忙。 15 01:25:59.698 31122-31122/com.my
我在 SO 上找到并阅读了各种帖子,其中提到您应该使用 getApplicationContext()绑定(bind)到 Service 时- 而不是 this (在 Activity 中)或 get
我正在开发绑定(bind)到本地服务的 Activity (在 Activity 的 onCreate 中): bindService(new Intent(this, CommandService.
我有一个与此相关的问题 question @mnish 大约一年前问过这个问题。 请看一下他的问题和代码。他实现了一个 ServiceConnection() 并将其传递给 bindService()
我有几个 Android Service,我想在我的 Activity 中绑定(bind)它们,因此我可以监视用户的多个操作。 为了能够绑定(bind)每个服务,我将有多个服务,我是否需要在我的 Ac
我正在使用启动 IntentService 的 BraodCastReceiver。一切看起来都很好,但我收到这个错误,我不知道它的来源: android.app.ServiceConnectionL
我有一个在启动时使用文字转语音的应用程序。一切似乎都运行良好,运行该应用程序时我没有遇到任何问题。但是,每次应用程序启动时,我都会收到一个 LogCat 错误,提示我的文本到语音转换中存在泄漏的 Se
我有一个非常简单的 Activity : public class MainActivity extends Activity { private Intent serviceInt
我正在使用 AIDL 在客户端 Activity 和服务之间执行 IPC。 ServiceConnection.onServiceConnected() 似乎在使用 bindService() 绑定(
我让这个应用程序监听传入的消息并大声朗读它们。问题是我退出时出现以下错误 12-21 15:45:29.949: E/ActivityThread(566): Activity mo.rach.col
能否请您向我解释一下,当我们绑定(bind)到服务但从不启动它然后解除绑定(bind)时会发生什么?我收到“Activity 泄露了一个服务连接错误”,但我不明白为什么。 我的服务: 包 com.ex
我在这里查看 Android 应用内结算教程: http://developer.android.com/guide/google/play/billing/billing_integrate.htm
我是一名优秀的程序员,十分优秀!