期刊文献+
共找到160篇文章
< 1 2 8 >
每页显示 20 50 100
An Empirical Study on the Effectiveness of Adversarial Examples in Malware Detection
1
作者 Younghoon Ban Myeonghyun Kim Haehyun Cho 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第6期3535-3563,共29页
Antivirus vendors and the research community employ Machine Learning(ML)or Deep Learning(DL)-based static analysis techniques for efficient identification of new threats,given the continual emergence of novel malware ... Antivirus vendors and the research community employ Machine Learning(ML)or Deep Learning(DL)-based static analysis techniques for efficient identification of new threats,given the continual emergence of novel malware variants.On the other hand,numerous researchers have reported that Adversarial Examples(AEs),generated by manipulating previously detected malware,can successfully evade ML/DL-based classifiers.Commercial antivirus systems,in particular,have been identified as vulnerable to such AEs.This paper firstly focuses on conducting black-box attacks to circumvent ML/DL-based malware classifiers.Our attack method utilizes seven different perturbations,including Overlay Append,Section Append,and Break Checksum,capitalizing on the ambiguities present in the PE format,as previously employed in evasion attack research.By directly applying the perturbation techniques to PE binaries,our attack method eliminates the need to grapple with the problem-feature space dilemma,a persistent challenge in many evasion attack studies.Being a black-box attack,our method can generate AEs that successfully evade both DL-based and ML-based classifiers.Also,AEs generated by the attack method retain their executability and malicious behavior,eliminating the need for functionality verification.Through thorogh evaluations,we confirmed that the attack method achieves an evasion rate of 65.6%against well-known ML-based malware detectors and can reach a remarkable 99%evasion rate against well-known DL-based malware detectors.Furthermore,our AEs demonstrated the capability to bypass detection by 17%of vendors out of the 64 on VirusTotal(VT).In addition,we propose a defensive approach that utilizes Trend Locality Sensitive Hashing(TLSH)to construct a similarity-based defense model.Through several experiments on the approach,we verified that our defense model can effectively counter AEs generated by the perturbation techniques.In conclusion,our defense model alleviates the limitation of the most promising defense method,adversarial training,which is only effective against the AEs that are included in the training classifiers. 展开更多
关键词 Malware classification machine learning adversarial examples evasion attack CYBERSECURITY
下载PDF
Omni-Detection of Adversarial Examples with Diverse Magnitudes
2
作者 Ke Jianpeng Wang Wenqi +3 位作者 Yang Kang Wang Lina Ye Aoshuang Wang Run 《China Communications》 SCIE CSCD 2024年第12期139-151,共13页
Deep neural networks(DNNs)are poten-tially susceptible to adversarial examples that are ma-liciously manipulated by adding imperceptible pertur-bations to legitimate inputs,leading to abnormal be-havior of models.Plen... Deep neural networks(DNNs)are poten-tially susceptible to adversarial examples that are ma-liciously manipulated by adding imperceptible pertur-bations to legitimate inputs,leading to abnormal be-havior of models.Plenty of methods have been pro-posed to defend against adversarial examples.How-ever,the majority of them are suffering the follow-ing weaknesses:1)lack of generalization and prac-ticality.2)fail to deal with unknown attacks.To ad-dress the above issues,we design the adversarial na-ture eraser(ANE)and feature map detector(FMD)to detect fragile and high-intensity adversarial examples,respectively.Then,we apply the ensemble learning method to compose our detector,dealing with adver-sarial examples with diverse magnitudes in a divide-and-conquer manner.Experimental results show that our approach achieves 99.30%and 99.62%Area un-der Curve(AUC)scores on average when tested with various Lp norm-based attacks on CIFAR-10 and Im-ageNet,respectively.Furthermore,our approach also shows its potential in detecting unknown attacks. 展开更多
关键词 adversarial example detection ensemble learning feature maps fragile and high-intensity ad-versarial examples
下载PDF
Auto-expanded multi query examples technology in content-based image retrieval 被引量:1
3
作者 王小玲 谢康林 《Journal of Southeast University(English Edition)》 EI CAS 2005年第3期287-292,共6页
In order to narrow the semantic gap existing in content-based image retrieval (CBIR),a novel retrieval technology called auto-extended multi query examples (AMQE) is proposed.It expands the single one query image ... In order to narrow the semantic gap existing in content-based image retrieval (CBIR),a novel retrieval technology called auto-extended multi query examples (AMQE) is proposed.It expands the single one query image used in traditional image retrieval into multi query examples so as to include more image features related with semantics.Retrieving images for each of the multi query examples and integrating the retrieval results,more relevant images can be obtained.The property of the recall-precision curve of a general retrieval algorithm and the K-means clustering method are used to realize the expansion according to the distance of image features of the initially retrieved images.The experimental results demonstrate that the AMQE technology can greatly improve the recall and precision of the original algorithms. 展开更多
关键词 content-based image retrieval SEMANTIC multi query examples K-means clustering
下载PDF
Patterns of Clay Minerals Transformation in Clay Gouge, with Examples from Revers Fault Rocks in Devonina Niqiuhe Formation in The Dayangshu Basin 被引量:2
4
作者 MENG Jie LI Benxian +1 位作者 ZHANG Juncheng LIU Xiaoyang 《Acta Geologica Sinica(English Edition)》 SCIE CAS CSCD 2017年第S1期59-60,共2页
The role of authigenic clay growth in clay gouge is increasingly recognized as a key to understanding the mechanics of berittle faulting and fault zone processes,including creep and seismogenesis,and providing new ins... The role of authigenic clay growth in clay gouge is increasingly recognized as a key to understanding the mechanics of berittle faulting and fault zone processes,including creep and seismogenesis,and providing new insights into the ongoing debate about the frictional strength of brittle fault(Haines and van der Pluijm,2012).However,neither the conditions nor the processes which 展开更多
关键词 with examples from Revers Fault Rocks in Devonina Niqiuhe Formation in The Dayangshu Basin Patterns of Clay Minerals Transformation in Clay Gouge
下载PDF
A Survey on Adversarial Examples in Deep Learning 被引量:2
5
作者 Kai Chen Haoqi Zhu +1 位作者 Leiming Yan Jinwei Wang 《Journal on Big Data》 2020年第2期71-84,共14页
Adversarial examples are hot topics in the field of security in deep learning.The feature,generation methods,attack and defense methods of the adversarial examples are focuses of the current research on adversarial ex... Adversarial examples are hot topics in the field of security in deep learning.The feature,generation methods,attack and defense methods of the adversarial examples are focuses of the current research on adversarial examples.This article explains the key technologies and theories of adversarial examples from the concept of adversarial examples,the occurrences of the adversarial examples,the attacking methods of adversarial examples.This article lists the possible reasons for the adversarial examples.This article also analyzes several typical generation methods of adversarial examples in detail:Limited-memory BFGS(L-BFGS),Fast Gradient Sign Method(FGSM),Basic Iterative Method(BIM),Iterative Least-likely Class Method(LLC),etc.Furthermore,in the perspective of the attack methods and reasons of the adversarial examples,the main defense techniques for the adversarial examples are listed:preprocessing,regularization and adversarial training method,distillation method,etc.,which application scenarios and deficiencies of different defense measures are pointed out.This article further discusses the application of adversarial examples which currently is mainly used in adversarial evaluation and adversarial training.Finally,the overall research direction of the adversarial examples is prospected to completely solve the adversarial attack problem.There are still a lot of practical and theoretical problems that need to be solved.Finding out the characteristics of the adversarial examples,giving a mathematical description of its practical application prospects,exploring the universal method of adversarial example generation and the generation mechanism of the adversarial examples are the main research directions of the adversarial examples in the future. 展开更多
关键词 Adversarial examples generation methods defense methods
下载PDF
A new method of constructing adversarial examplesfor quantum variational circuits
6
作者 颜金歌 闫丽丽 张仕斌 《Chinese Physics B》 SCIE EI CAS CSCD 2023年第7期268-272,共5页
A quantum variational circuit is a quantum machine learning model similar to a neural network.A crafted adversarial example can lead to incorrect results for the model.Using adversarial examples to train the model wil... A quantum variational circuit is a quantum machine learning model similar to a neural network.A crafted adversarial example can lead to incorrect results for the model.Using adversarial examples to train the model will greatly improve its robustness.The existing method is to use automatic differentials or finite difference to obtain a gradient and use it to construct adversarial examples.This paper proposes an innovative method for constructing adversarial examples of quantum variational circuits.In this method,the gradient can be obtained by measuring the expected value of a quantum bit respectively in a series quantum circuit.This method can be used to construct the adversarial examples for a quantum variational circuit classifier.The implementation results prove the effectiveness of the proposed method.Compared with the existing method,our method requires fewer resources and is more efficient. 展开更多
关键词 quantum variational circuit adversarial examples quantum machine learning quantum circuit
下载PDF
Examples for Clinical Use of Ma Zi Ren Wan
7
作者 张书文 段树民 《Journal of Traditional Chinese Medicine》 SCIE CAS CSCD 2002年第3期216-217,共2页
  Ma Zi Ren Wan (麻子仁丸), originally recorded in Treatise on Febrile Diseases (伤寒论), is composed of Ma Zi Ren (麻子仁Fructus Cannabis), Bai Shao (白芍Radix Paeoniae Alba), Zhi Shi (枳实Fructus Aurantii Immaturu...   Ma Zi Ren Wan (麻子仁丸), originally recorded in Treatise on Febrile Diseases (伤寒论), is composed of Ma Zi Ren (麻子仁Fructus Cannabis), Bai Shao (白芍Radix Paeoniae Alba), Zhi Shi (枳实Fructus Aurantii Immaturus), Da Huang (大黄Radix etRhizoma Rhei), Hou Po (厚朴cortex Magnoliae Officinalis) and Xing Ren (杏仁Semen Armeniacae Amarum). Good therapeutic results have been achieved by using Ma ZiRen Wan in treatment of febrile disease at the restoring stage, chronic consumptive diseases, hemorrhoid, disorders in women after delivery, chronic kidney disease, senile constipation, pulmonary heart disease, diabetes, coronary heart disease and hypertension. Some illustrative cases are introduced below.   …… 展开更多
关键词 麻子仁 火麻仁 examples for Clinical Use of Ma Zi Ren Wan 厚朴 木兰属
下载PDF
Defending Adversarial Examples by a Clipped Residual U-Net Model
8
作者 Kazim Ali Adnan N.Qureshi +2 位作者 Muhammad Shahid Bhatti Abid Sohail Mohammad Hijji 《Intelligent Automation & Soft Computing》 SCIE 2023年第2期2237-2256,共20页
Deep learning-based systems have succeeded in many computer vision tasks.However,it is found that the latest study indicates that these systems are in danger in the presence of adversarial attacks.These attacks can qu... Deep learning-based systems have succeeded in many computer vision tasks.However,it is found that the latest study indicates that these systems are in danger in the presence of adversarial attacks.These attacks can quickly spoil deep learning models,e.g.,different convolutional neural networks(CNNs),used in various computer vision tasks from image classification to object detection.The adversarial examples are carefully designed by injecting a slight perturbation into the clean images.The proposed CRU-Net defense model is inspired by state-of-the-art defense mechanisms such as MagNet defense,Generative Adversarial Net-work Defense,Deep Regret Analytic Generative Adversarial Networks Defense,Deep Denoising Sparse Autoencoder Defense,and Condtional Generattive Adversarial Network Defense.We have experimentally proved that our approach is better than previous defensive techniques.Our proposed CRU-Net model maps the adversarial image examples into clean images by eliminating the adversarial perturbation.The proposed defensive approach is based on residual and U-Net learning.Many experiments are done on the datasets MNIST and CIFAR10 to prove that our proposed CRU-Net defense model prevents adversarial example attacks in WhiteBox and BlackBox settings and improves the robustness of the deep learning algorithms especially in the computer visionfield.We have also reported similarity(SSIM and PSNR)between the original and restored clean image examples by the proposed CRU-Net defense model. 展开更多
关键词 Adversarial examples adversarial attacks defense method residual learning u-net cgan cru-et model
下载PDF
The algorithm AE_(11) of learning from examples
9
作者 ZHANG Hai-yi BI Jian-dong 《Journal of Harbin Institute of Technology(New Series)》 EI CAS 2006年第2期226-232,共7页
We first put forward the idea of a positive extension matrix (PEM) on paper. Then, an algorithm, AE_ 11, was built with the aid of the PEM. Finally, we made the comparisons of our experimental results and the final re... We first put forward the idea of a positive extension matrix (PEM) on paper. Then, an algorithm, AE_ 11, was built with the aid of the PEM. Finally, we made the comparisons of our experimental results and the final result was fairly satisfying. 展开更多
关键词 learning from examples concept acquisition inductive learning knowledge acquisition
下载PDF
Environmental Implication of Subaqueous Lava Flows from A Continental Large Igneous Province: Examples from the Moroccan Central Atlantic Magmatic Province(CAMP)
10
作者 S.EL GHILANI N.YOUBI +4 位作者 J.MADEIRA E.H.CHELLAI ALBERTO LóPEZGALINDO L.MARTINS J.MATA 《Acta Geologica Sinica(English Edition)》 SCIE CAS CSCD 2016年第S1期117-,共1页
The Early Jurassic volcanic sequence of the Central Atlantic Magmatic Province(CAMP)of Morocco is classically subdivided into four stratigraphic units:the Lower,Middle,Upper and Recurrent Formations separated
关键词 examples from the Moroccan Central Atlantic Magmatic Province Environmental Implication of Subaqueous Lava Flows from A Continental Large Igneous Province CAMP
下载PDF
SOME EXAMPLES OF TRANSITION SPANS MINERALIZING AND ANALYSIS OFTHEIR DYNAMICS
11
作者 BI Hua1, ZHAO Zhi zhong1, ZHU Wei huang2, YANG Yuan gen2,HUANG Lan3 and LIU Qiang3(1. Department of Resources, Environment and Tourism, Hainan Normal College, Haikou 571158, China 2. Institute of Geochemistry, Chinese Academy of Sciences, Guiyang 550002, China 3. Biology Department, Hainan Normal College,Haikou 571158,China) 《Geotectonica et Metallogenia》 2002年第1期89-92,共4页
The transitional span is a special environment for deposits. Taking peat, oil gas, metallic deposits as examples, this paper discusses the spatial temporal transitional characteristics of mineralization in transitiona... The transitional span is a special environment for deposits. Taking peat, oil gas, metallic deposits as examples, this paper discusses the spatial temporal transitional characteristics of mineralization in transitional regions, points out the importance of the mineralization in transition spans, and analyses their dynamics finally. 展开更多
关键词 TRANSITION SPANS ORE formation examples DYNAMICS
下载PDF
Examples of Folklor in Dainba
12
作者 NGAZHISDANBO LHALOGYIA 《China's Tibet》 1997年第6期43-43,共1页
关键词 examples of Folklor in Dainba
下载PDF
An Intelligent Secure Adversarial Examples Detection Scheme in Heterogeneous Complex Environments
13
作者 Weizheng Wang Xiangqi Wang +5 位作者 Xianmin Pan Xingxing Gong Jian Liang Pradip Kumar Sharma Osama Alfarraj Wael Said 《Computers, Materials & Continua》 SCIE EI 2023年第9期3859-3876,共18页
Image-denoising techniques are widely used to defend against Adversarial Examples(AEs).However,denoising alone cannot completely eliminate adversarial perturbations.The remaining perturbations tend to amplify as they ... Image-denoising techniques are widely used to defend against Adversarial Examples(AEs).However,denoising alone cannot completely eliminate adversarial perturbations.The remaining perturbations tend to amplify as they propagate through deeper layers of the network,leading to misclassifications.Moreover,image denoising compromises the classification accuracy of original examples.To address these challenges in AE defense through image denoising,this paper proposes a novel AE detection technique.The proposed technique combines multiple traditional image-denoising algorithms and Convolutional Neural Network(CNN)network structures.The used detector model integrates the classification results of different models as the input to the detector and calculates the final output of the detector based on a machine-learning voting algorithm.By analyzing the discrepancy between predictions made by the model on original examples and denoised examples,AEs are detected effectively.This technique reduces computational overhead without modifying the model structure or parameters,effectively avoiding the error amplification caused by denoising.The proposed approach demonstrates excellent detection performance against mainstream AE attacks.Experimental results show outstanding detection performance in well-known AE attacks,including Fast Gradient Sign Method(FGSM),Basic Iteration Method(BIM),DeepFool,and Carlini&Wagner(C&W),achieving a 94%success rate in FGSM detection,while only reducing the accuracy of clean examples by 4%. 展开更多
关键词 Deep neural networks adversarial example image denoising adversarial example detection machine learning adversarial attack
下载PDF
Adversarial Examples Protect Your Privacy on Speech Enhancement System
14
作者 Mingyu Dong Diqun Yan Rangding Wang 《Computer Systems Science & Engineering》 SCIE EI 2023年第7期1-12,共12页
Speech is easily leaked imperceptibly.When people use their phones,the personal voice assistant is constantly listening and waiting to be activated.Private content in speech may be maliciously extracted through automa... Speech is easily leaked imperceptibly.When people use their phones,the personal voice assistant is constantly listening and waiting to be activated.Private content in speech may be maliciously extracted through automatic speech recognition(ASR)technology by some applications on phone devices.To guarantee that the recognized speech content is accurate,speech enhancement technology is used to denoise the input speech.Speech enhancement technology has developed rapidly along with deep neural networks(DNNs),but adversarial examples can cause DNNs to fail.Considering that the vulnerability of DNN can be used to protect the privacy in speech.In this work,we propose an adversarial method to degrade speech enhancement systems,which can prevent the malicious extraction of private information in speech.Experimental results show that the generated enhanced adversarial examples can be removed most content of the target speech or replaced with target speech content by speech enhancement.The word error rate(WER)between the enhanced original example and enhanced adversarial example recognition result can reach 89.0%.WER of target attack between enhanced adversarial example and target example is low at 33.75%.The adversarial perturbation in the adversarial example can bring much more change than itself.The rate of difference between two enhanced examples and adversarial perturbation can reach more than 1.4430.Meanwhile,the transferability between different speech enhancement models is also investigated.The low transferability of the method can be used to ensure the content in the adversarial example is not damaged,the useful information can be extracted by the friendly ASR.This work can prevent the malicious extraction of speech. 展开更多
关键词 Adversarial example speech enhancement privacy protection deep neural network
下载PDF
Eye Movement Modeling Examples as a Teaching Tool for Epidural Block
15
作者 Emanuele Capogna Marco Scorzoni +5 位作者 Gian Luigi Gonnella Matteo Velardo Pier Paolo Giurì Mariano Ciancia Gaetano Draisci Giorgio Capogna 《Open Journal of Anesthesiology》 2023年第4期75-84,共10页
Introduction: Video examples with task demonstrations by experts, with the expert’s eye movements superimposed on the task, are known as “eye movement modeling examples” (EMME). We performed this study to evaluate ... Introduction: Video examples with task demonstrations by experts, with the expert’s eye movements superimposed on the task, are known as “eye movement modeling examples” (EMME). We performed this study to evaluate if there were improvements in the performance of anesthesia novice trainees when executing the epidural technique after an EMME of epidural block procedure. Methods: We developed an eye movement modeling example (EMME) from eye tracking recordings made by experienced anesthesiologists with more than 20 years of experience. Forty-two PGY3 anesthesia trainees who had never previously performed an epidural block were randomized to receive (study group) or not receive (control group) the EMME video before their institutional training. All the trainees were evaluated every 10 epidural blocks until the end of the rotation period, by an independent, blinded observer using the Global Rating Scale for Epidural Anesthesia (GRS). Results: Trainees who received the EMME training exhibited more respect for the patient’s tissues (P Discussion: This is the first study that has used the EMME for a practical, clinical teaching purpose on real patients and that has used it as an aid in teaching epidural anesthesia. We demonstrated that inexperienced trainees who received the EMME training improved their proficiency at epidural blocks as compared to those who had no EMME training beforehand. Given this result, we welcome further studies to investigate the impact and the role of EMME on clinical teaching in the field of anesthesia. 展开更多
关键词 Eye Tracking Eye Movement Modeling Example Epidural Anesthesia
下载PDF
On some examples of River Capture
16
作者 C.Y.Hsieh 《地理学报》 EI 1935年第3期152-153,共2页
In the course of stream development, each stream and its tributaries tend toincrease their length by headward erosion, or in other words to enlarge their catchment area. A shifting of divide and eventually a struggle ... In the course of stream development, each stream and its tributaries tend toincrease their length by headward erosion, or in other words to enlarge their catchment area. A shifting of divide and eventually a struggle for existence by which a part or the whole of tne stream will be diverted or captured by the otherwill be resulted. This phenomena is called in physiography, river capture or river piracy. 展开更多
关键词 On some examples of River Capture
下载PDF
Exploratory Research on Defense against Natural Adversarial Examples in Image Classification
17
作者 Yaoxuan Zhu Hua Yang Bin Zhu 《Computers, Materials & Continua》 2025年第2期1947-1968,共22页
The emergence of adversarial examples has revealed the inadequacies in the robustness of image classification models based on Convolutional Neural Networks (CNNs). Particularly in recent years, the discovery of natura... The emergence of adversarial examples has revealed the inadequacies in the robustness of image classification models based on Convolutional Neural Networks (CNNs). Particularly in recent years, the discovery of natural adversarial examples has posed significant challenges, as traditional defense methods against adversarial attacks have proven to be largely ineffective against these natural adversarial examples. This paper explores defenses against these natural adversarial examples from three perspectives: adversarial examples, model architecture, and dataset. First, it employs Class Activation Mapping (CAM) to visualize how models classify natural adversarial examples, identifying several typical attack patterns. Next, various common CNN models are analyzed to evaluate their susceptibility to these attacks, revealing that different architectures exhibit varying defensive capabilities. The study finds that as the depth of a network increases, its defenses against natural adversarial examples strengthen. Lastly, Finally, the impact of dataset class distribution on the defense capability of models is examined, focusing on two aspects: the number of classes in the training set and the number of predicted classes. This study investigates how these factors influence the model’s ability to defend against natural adversarial examples. Results indicate that reducing the number of training classes enhances the model’s defense against natural adversarial examples. Additionally, under a fixed number of training classes, some CNN models show an optimal range of predicted classes for achieving the best defense performance against these adversarial examples. 展开更多
关键词 Image classification convolutional neural network natural adversarial example data set defense against adversarial examples
下载PDF
Low-rank matrix recovery with total generalized variation for defending adversarial examples
18
作者 Wen LI Hengyou WANG +4 位作者 Lianzhi HUO Qiang HE Linlin CHEN Zhiquan HE Wing W.Y.Ng 《Frontiers of Information Technology & Electronic Engineering》 SCIE EI CSCD 2024年第3期432-445,共14页
Low-rank matrix decomposition with first-order total variation(TV)regularization exhibits excellent performance in exploration of image structure.Taking advantage of its excellent performance in image denoising,we app... Low-rank matrix decomposition with first-order total variation(TV)regularization exhibits excellent performance in exploration of image structure.Taking advantage of its excellent performance in image denoising,we apply it to improve the robustness of deep neural networks.However,although TV regularization can improve the robustness of the model,it reduces the accuracy of normal samples due to its over-smoothing.In our work,we develop a new low-rank matrix recovery model,called LRTGV,which incorporates total generalized variation(TGV)regularization into the reweighted low-rank matrix recovery model.In the proposed model,TGV is used to better reconstruct texture information without over-smoothing.The reweighted nuclear norm and Li-norm can enhance the global structure information.Thus,the proposed LRTGV can destroy the structure of adversarial noise while re-enhancing the global structure and local texture of the image.To solve the challenging optimal model issue,we propose an algorithm based on the alternating direction method of multipliers.Experimental results show that the proposed algorithm has a certain defense capability against black-box attacks,and outperforms state-of-the-art low-rank matrix recovery methods in image restoration. 展开更多
关键词 Total generalized variation Low-rank matrix Alternating direction method of multipliers Adversarial example
原文传递
PROVING-BY-EXAMPLES METHOD AND INCLUSION OF VARIETIES
19
作者 王东明 《Chinese Science Bulletin》 SCIE EI CAS 1988年第24期2015-2018,共4页
Based on Ritt-Wu well ordering principle and Wu’s constructive theory of decomposing a polynomial set into irreducible ascending sets, we show that Hong’s 'Provingby Examples' method suits all theorems of eq... Based on Ritt-Wu well ordering principle and Wu’s constructive theory of decomposing a polynomial set into irreducible ascending sets, we show that Hong’s 'Provingby Examples' method suits all theorems of equation type. Moreover, the de- 展开更多
关键词 proving by examples Wu’s METHOD geometry THEOREMS of equation type ALGEBRAIC VARIETY
原文传递
EXPAT CONSUMERS IGNORED Even as successful examples of expat-oriented businesses abound, marketers couldn’t care less. Should they care more?
20
作者 MATT YOUNG 《Beijing Review》 2007年第1期36-37,共2页
关键词 Should they care more exp marketers couldn t care less EXPAT CONSUMERS IGNORED Even as successful examples of expat-oriented businesses abound
原文传递
上一页 1 2 8 下一页 到第
使用帮助 返回顶部