hadoop第二坑

#Hadoop#第二坑

  • 需求
    实现两张表的join
    order表
    -
    pd表
    报以下错误
D:\Java\JDK\bin\java.exe "-javaagent:D:\IntelliJ IDEA 2018.2.2\lib\idea_rt.jar=7031:D:\IntelliJ IDEA 2018.2.2\bin" -Dfile.encoding=UTF-8 -classpath D:\Java\JDK\jre\lib\charsets.jar;D:\Java\JDK\jre\lib\deploy.jar;D:\Java\JDK\jre\lib\ext\access-bridge-64.jar;D:\Java\JDK\jre\lib\ext\cldrdata.jar;D:\Java\JDK\jre\lib\ext\dnsns.jar;D:\Java\JDK\jre\lib\ext\jaccess.jar;D:\Java\JDK\jre\lib\ext\jfxrt.jar;D:\Java\JDK\jre\lib\ext\localedata.jar;D:\Java\JDK\jre\lib\ext\nashorn.jar;D:\Java\JDK\jre\lib\ext\sunec.jar;D:\Java\JDK\jre\lib\ext\sunjce_provider.jar;D:\Java\JDK\jre\lib\ext\sunmscapi.jar;D:\Java\JDK\jre\lib\ext\sunpkcs11.jar;D:\Java\JDK\jre\lib\ext\zipfs.jar;D:\Java\JDK\jre\lib\javaws.jar;D:\Java\JDK\jre\lib\jce.jar;D:\Java\JDK\jre\lib\jfr.jar;D:\Java\JDK\jre\lib\jfxswt.jar;D:\Java\JDK\jre\lib\jsse.jar;D:\Java\JDK\jre\lib\management-agent.jar;D:\Java\JDK\jre\lib\plugin.jar;D:\Java\JDK\jre\lib\resources.jar;D:\Java\JDK\jre\lib\rt.jar;C:\Users\Mr.Robot\Desktop\workspace\mapreduce\target\classes;C:\Users\Mr.Robot\.m2\repository\junit\junit\4.13-beta-1\junit-4.13-beta-1.jar;C:\Users\Mr.Robot\.m2\repository\org\hamcrest\hamcrest-core\1.3\hamcrest-core-1.3.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\logging\log4j\log4j-core\2.8.2\log4j-core-2.8.2.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\logging\log4j\log4j-api\2.8.2\log4j-api-2.8.2.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\hadoop\hadoop-common\2.7.2\hadoop-common-2.7.2.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\hadoop\hadoop-annotations\2.7.2\hadoop-annotations-2.7.2.jar;D:\Java\JDK\lib\tools.jar;C:\Users\Mr.Robot\.m2\repository\com\google\guava\guava\11.0.2\guava-11.0.2.jar;C:\Users\Mr.Robot\.m2\repository\commons-cli\commons-cli\1.2\commons-cli-1.2.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\commons\commons-math3\3.1.1\commons-math3-3.1.1.jar;C:\Users\Mr.Robot\.m2\repository\xmlenc\xmlenc\0.52\xmlenc-0.52.jar;C:\Users\Mr.Robot\.m2\repository\commons-httpclient\commons-httpclient\3.1\commons-httpclient-3.1.jar;C:\Users\Mr.Robot\.m2\repository\commons-codec\commons-codec\1.4\commons-codec-1.4.jar;C:\Users\Mr.Robot\.m2\repository\commons-io\commons-io\2.4\commons-io-2.4.jar;C:\Users\Mr.Robot\.m2\repository\commons-net\commons-net\3.1\commons-net-3.1.jar;C:\Users\Mr.Robot\.m2\repository\commons-collections\commons-collections\3.2.2\commons-collections-3.2.2.jar;C:\Users\Mr.Robot\.m2\repository\javax\servlet\servlet-api\2.5\servlet-api-2.5.jar;C:\Users\Mr.Robot\.m2\repository\org\mortbay\jetty\jetty\6.1.26\jetty-6.1.26.jar;C:\Users\Mr.Robot\.m2\repository\org\mortbay\jetty\jetty-util\6.1.26\jetty-util-6.1.26.jar;C:\Users\Mr.Robot\.m2\repository\javax\servlet\jsp\jsp-api\2.1\jsp-api-2.1.jar;C:\Users\Mr.Robot\.m2\repository\com\sun\jersey\jersey-core\1.9\jersey-core-1.9.jar;C:\Users\Mr.Robot\.m2\repository\com\sun\jersey\jersey-json\1.9\jersey-json-1.9.jar;C:\Users\Mr.Robot\.m2\repository\org\codehaus\jettison\jettison\1.1\jettison-1.1.jar;C:\Users\Mr.Robot\.m2\repository\com\sun\xml\bind\jaxb-impl\2.2.3-1\jaxb-impl-2.2.3-1.jar;C:\Users\Mr.Robot\.m2\repository\javax\xml\bind\jaxb-api\2.2.2\jaxb-api-2.2.2.jar;C:\Users\Mr.Robot\.m2\repository\javax\xml\stream\stax-api\1.0-2\stax-api-1.0-2.jar;C:\Users\Mr.Robot\.m2\repository\javax\activation\activation\1.1\activation-1.1.jar;C:\Users\Mr.Robot\.m2\repository\org\codehaus\jackson\jackson-jaxrs\1.8.3\jackson-jaxrs-1.8.3.jar;C:\Users\Mr.Robot\.m2\repository\org\codehaus\jackson\jackson-xc\1.8.3\jackson-xc-1.8.3.jar;C:\Users\Mr.Robot\.m2\repository\com\sun\jersey\jersey-server\1.9\jersey-server-1.9.jar;C:\Users\Mr.Robot\.m2\repository\asm\asm\3.1\asm-3.1.jar;C:\Users\Mr.Robot\.m2\repository\commons-logging\commons-logging\1.1.3\commons-logging-1.1.3.jar;C:\Users\Mr.Robot\.m2\repository\log4j\log4j\1.2.17\log4j-1.2.17.jar;C:\Users\Mr.Robot\.m2\repository\net\java\dev\jets3t\jets3t\0.9.0\jets3t-0.9.0.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\httpcomponents\httpclient\4.1.2\httpclient-4.1.2.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\httpcomponents\httpcore\4.1.2\httpcore-4.1.2.jar;C:\Users\Mr.Robot\.m2\repository\com\jamesmurty\utils\java-xmlbuilder\0.4\java-xmlbuilder-0.4.jar;C:\Users\Mr.Robot\.m2\repository\commons-lang\commons-lang\2.6\commons-lang-2.6.jar;C:\Users\Mr.Robot\.m2\repository\commons-configuration\commons-configuration\1.6\commons-configuration-1.6.jar;C:\Users\Mr.Robot\.m2\repository\commons-digester\commons-digester\1.8\commons-digester-1.8.jar;C:\Users\Mr.Robot\.m2\repository\commons-beanutils\commons-beanutils\1.7.0\commons-beanutils-1.7.0.jar;C:\Users\Mr.Robot\.m2\repository\commons-beanutils\commons-beanutils-core\1.8.0\commons-beanutils-core-1.8.0.jar;C:\Users\Mr.Robot\.m2\repository\org\slf4j\slf4j-api\1.7.10\slf4j-api-1.7.10.jar;C:\Users\Mr.Robot\.m2\repository\org\slf4j\slf4j-log4j12\1.7.10\slf4j-log4j12-1.7.10.jar;C:\Users\Mr.Robot\.m2\repository\org\codehaus\jackson\jackson-core-asl\1.9.13\jackson-core-asl-1.9.13.jar;C:\Users\Mr.Robot\.m2\repository\org\codehaus\jackson\jackson-mapper-asl\1.9.13\jackson-mapper-asl-1.9.13.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\avro\avro\1.7.4\avro-1.7.4.jar;C:\Users\Mr.Robot\.m2\repository\com\thoughtworks\paranamer\paranamer\2.3\paranamer-2.3.jar;C:\Users\Mr.Robot\.m2\repository\org\xerial\snappy\snappy-java\1.0.4.1\snappy-java-1.0.4.1.jar;C:\Users\Mr.Robot\.m2\repository\com\google\protobuf\protobuf-java\2.5.0\protobuf-java-2.5.0.jar;C:\Users\Mr.Robot\.m2\repository\com\google\code\gson\gson\2.2.4\gson-2.2.4.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\hadoop\hadoop-auth\2.7.2\hadoop-auth-2.7.2.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\directory\server\apacheds-kerberos-codec\2.0.0-M15\apacheds-kerberos-codec-2.0.0-M15.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\directory\server\apacheds-i18n\2.0.0-M15\apacheds-i18n-2.0.0-M15.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\directory\api\api-asn1-api\1.0.0-M20\api-asn1-api-1.0.0-M20.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\directory\api\api-util\1.0.0-M20\api-util-1.0.0-M20.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\curator\curator-framework\2.7.1\curator-framework-2.7.1.jar;C:\Users\Mr.Robot\.m2\repository\com\jcraft\jsch\0.1.42\jsch-0.1.42.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\curator\curator-client\2.7.1\curator-client-2.7.1.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\curator\curator-recipes\2.7.1\curator-recipes-2.7.1.jar;C:\Users\Mr.Robot\.m2\repository\com\google\code\findbugs\jsr305\3.0.0\jsr305-3.0.0.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\htrace\htrace-core\3.1.0-incubating\htrace-core-3.1.0-incubating.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\zookeeper\zookeeper\3.4.6\zookeeper-3.4.6.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\commons\commons-compress\1.4.1\commons-compress-1.4.1.jar;C:\Users\Mr.Robot\.m2\repository\org\tukaani\xz\1.0\xz-1.0.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\hadoop\hadoop-client\2.7.2\hadoop-client-2.7.2.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\hadoop\hadoop-mapreduce-client-app\2.7.2\hadoop-mapreduce-client-app-2.7.2.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\hadoop\hadoop-mapreduce-client-common\2.7.2\hadoop-mapreduce-client-common-2.7.2.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\hadoop\hadoop-yarn-client\2.7.2\hadoop-yarn-client-2.7.2.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\hadoop\hadoop-yarn-server-common\2.7.2\hadoop-yarn-server-common-2.7.2.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\hadoop\hadoop-mapreduce-client-shuffle\2.7.2\hadoop-mapreduce-client-shuffle-2.7.2.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\hadoop\hadoop-yarn-api\2.7.2\hadoop-yarn-api-2.7.2.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\hadoop\hadoop-mapreduce-client-core\2.7.2\hadoop-mapreduce-client-core-2.7.2.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\hadoop\hadoop-yarn-common\2.7.2\hadoop-yarn-common-2.7.2.jar;C:\Users\Mr.Robot\.m2\repository\com\sun\jersey\jersey-client\1.9\jersey-client-1.9.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\hadoop\hadoop-mapreduce-client-jobclient\2.7.2\hadoop-mapreduce-client-jobclient-2.7.2.jar;C:\Users\Mr.Robot\.m2\repository\org\apache\hadoop\hadoop-hdfs\2.7.2\hadoop-hdfs-2.7.2.jar;C:\Users\Mr.Robot\.m2\repository\commons-daemon\commons-daemon\1.0.13\commons-daemon-1.0.13.jar;C:\Users\Mr.Robot\.m2\repository\io\netty\netty\3.6.2.Final\netty-3.6.2.Final.jar;C:\Users\Mr.Robot\.m2\repository\io\netty\netty-all\4.0.23.Final\netty-all-4.0.23.Final.jar;C:\Users\Mr.Robot\.m2\repository\xerces\xercesImpl\2.9.1\xercesImpl-2.9.1.jar;C:\Users\Mr.Robot\.m2\repository\xml-apis\xml-apis\1.3.04\xml-apis-1.3.04.jar;C:\Users\Mr.Robot\.m2\repository\org\fusesource\leveldbjni\leveldbjni-all\1.8\leveldbjni-all-1.8.jar com.mrrobot.mr.table.TableDriver
2019-01-30 14:34:01,232 INFO [org.apache.hadoop.conf.Configuration.deprecation] - session.id is deprecated. Instead, use dfs.metrics.session-id
2019-01-30 14:34:01,233 INFO [org.apache.hadoop.metrics.jvm.JvmMetrics] - Initializing JVM Metrics with processName=JobTracker, sessionId=
2019-01-30 14:34:02,261 WARN [org.apache.hadoop.mapreduce.JobResourceUploader] - Hadoop command-line option parsing not performed. Implement the Tool interface and execute your application with ToolRunner to remedy this.
2019-01-30 14:34:02,288 WARN [org.apache.hadoop.mapreduce.JobResourceUploader] - No job jar file set.  User classes may not be found. See Job or Job#setJar(String).
2019-01-30 14:34:02,376 INFO [org.apache.hadoop.mapreduce.lib.input.FileInputFormat] - Total input paths to process : 2
2019-01-30 14:34:02,412 INFO [org.apache.hadoop.mapreduce.JobSubmitter] - number of splits:2
2019-01-30 14:34:02,502 INFO [org.apache.hadoop.mapreduce.JobSubmitter] - Submitting tokens for job: job_local1977368438_0001
2019-01-30 14:34:02,683 INFO [org.apache.hadoop.mapreduce.Job] - The url to track the job: http://localhost:8080/
2019-01-30 14:34:02,684 INFO [org.apache.hadoop.mapreduce.Job] - Running job: job_local1977368438_0001
2019-01-30 14:34:02,685 INFO [org.apache.hadoop.mapred.LocalJobRunner] - OutputCommitter set in config null
2019-01-30 14:34:02,690 INFO [org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter] - File Output Committer Algorithm version is 1
2019-01-30 14:34:02,693 INFO [org.apache.hadoop.mapred.LocalJobRunner] - OutputCommitter is org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
2019-01-30 14:34:02,784 INFO [org.apache.hadoop.mapred.LocalJobRunner] - Waiting for map tasks
2019-01-30 14:34:02,785 INFO [org.apache.hadoop.mapred.LocalJobRunner] - Starting task: attempt_local1977368438_0001_m_000000_0
2019-01-30 14:34:02,817 INFO [org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter] - File Output Committer Algorithm version is 1
2019-01-30 14:34:02,824 INFO [org.apache.hadoop.yarn.util.ProcfsBasedProcessTree] - ProcfsBasedProcessTree currently is supported only on Linux.
2019-01-30 14:34:02,862 INFO [org.apache.hadoop.mapred.Task] -  Using ResourceCalculatorProcessTree : org.apache.hadoop.yarn.util.WindowsBasedProcessTree@3aafe33c
2019-01-30 14:34:02,867 INFO [org.apache.hadoop.mapred.MapTask] - Processing split: file:/G:/大数据/尚硅谷/尚硅谷大数据()/input/inputTable/order.txt:0+64
2019-01-30 14:34:02,929 INFO [org.apache.hadoop.mapred.MapTask] - (EQUATOR) 0 kvi 26214396(104857584)
2019-01-30 14:34:02,929 INFO [org.apache.hadoop.mapred.MapTask] - mapreduce.task.io.sort.mb: 100
2019-01-30 14:34:02,929 INFO [org.apache.hadoop.mapred.MapTask] - soft limit at 83886080
2019-01-30 14:34:02,929 INFO [org.apache.hadoop.mapred.MapTask] - bufstart = 0; bufvoid = 104857600
2019-01-30 14:34:02,929 INFO [org.apache.hadoop.mapred.MapTask] - kvstart = 26214396; length = 6553600
2019-01-30 14:34:02,930 WARN [org.apache.hadoop.mapred.MapTask] - Unable to initialize MapOutputCollector org.apache.hadoop.mapred.MapTask$MapOutputBuffer
java.lang.NullPointerException
	at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.init(MapTask.java:1011)
	at org.apache.hadoop.mapred.MapTask.createSortingCollector(MapTask.java:402)
	at org.apache.hadoop.mapred.MapTask.access$100(MapTask.java:81)
	at org.apache.hadoop.mapred.MapTask$NewOutputCollector.<init>(MapTask.java:698)
	at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:770)
	at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)
	at org.apache.hadoop.mapred.LocalJobRunner$Job$MapTaskRunnable.run(LocalJobRunner.java:243)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
2019-01-30 14:34:02,932 INFO [org.apache.hadoop.mapred.LocalJobRunner] - Starting task: attempt_local1977368438_0001_m_000001_0
2019-01-30 14:34:02,933 INFO [org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter] - File Output Committer Algorithm version is 1
2019-01-30 14:34:02,933 INFO [org.apache.hadoop.yarn.util.ProcfsBasedProcessTree] - ProcfsBasedProcessTree currently is supported only on Linux.
2019-01-30 14:34:02,969 INFO [org.apache.hadoop.mapred.Task] -  Using ResourceCalculatorProcessTree : org.apache.hadoop.yarn.util.WindowsBasedProcessTree@2eaba032
2019-01-30 14:34:02,971 INFO [org.apache.hadoop.mapred.MapTask] - Processing split: file:/G:/大数据/尚硅谷/尚硅谷大数据()/input/inputTable/pd.txt:0+26
2019-01-30 14:34:03,025 INFO [org.apache.hadoop.mapred.MapTask] - (EQUATOR) 0 kvi 26214396(104857584)
2019-01-30 14:34:03,025 INFO [org.apache.hadoop.mapred.MapTask] - mapreduce.task.io.sort.mb: 100
2019-01-30 14:34:03,025 INFO [org.apache.hadoop.mapred.MapTask] - soft limit at 83886080
2019-01-30 14:34:03,025 INFO [org.apache.hadoop.mapred.MapTask] - bufstart = 0; bufvoid = 104857600
2019-01-30 14:34:03,025 INFO [org.apache.hadoop.mapred.MapTask] - kvstart = 26214396; length = 6553600
2019-01-30 14:34:03,025 WARN [org.apache.hadoop.mapred.MapTask] - Unable to initialize MapOutputCollector org.apache.hadoop.mapred.MapTask$MapOutputBuffer
java.lang.NullPointerException
	at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.init(MapTask.java:1011)
	at org.apache.hadoop.mapred.MapTask.createSortingCollector(MapTask.java:402)
	at org.apache.hadoop.mapred.MapTask.access$100(MapTask.java:81)
	at org.apache.hadoop.mapred.MapTask$NewOutputCollector.<init>(MapTask.java:698)
	at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:770)
	at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)
	at org.apache.hadoop.mapred.LocalJobRunner$Job$MapTaskRunnable.run(LocalJobRunner.java:243)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
2019-01-30 14:34:03,025 INFO [org.apache.hadoop.mapred.LocalJobRunner] - map task executor complete.
2019-01-30 14:34:03,027 WARN [org.apache.hadoop.mapred.LocalJobRunner] - job_local1977368438_0001
java.lang.Exception: java.io.IOException: Initialization of all the collectors failed. Error in last collector was :null
	at org.apache.hadoop.mapred.LocalJobRunner$Job.runTasks(LocalJobRunner.java:462)
	at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:522)
Caused by: java.io.IOException: Initialization of all the collectors failed. Error in last collector was :null
	at org.apache.hadoop.mapred.MapTask.createSortingCollector(MapTask.java:415)
	at org.apache.hadoop.mapred.MapTask.access$100(MapTask.java:81)
	at org.apache.hadoop.mapred.MapTask$NewOutputCollector.<init>(MapTask.java:698)
	at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:770)
	at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)
	at org.apache.hadoop.mapred.LocalJobRunner$Job$MapTaskRunnable.run(LocalJobRunner.java:243)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.NullPointerException
	at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.init(MapTask.java:1011)
	at org.apache.hadoop.mapred.MapTask.createSortingCollector(MapTask.java:402)
	... 10 more
2019-01-30 14:34:03,686 INFO [org.apache.hadoop.mapreduce.Job] - Job job_local1977368438_0001 running in uber mode : false
2019-01-30 14:34:03,687 INFO [org.apache.hadoop.mapreduce.Job] -  map 0% reduce 0%
2019-01-30 14:34:03,688 INFO [org.apache.hadoop.mapreduce.Job] - Job job_local1977368438_0001 failed with state FAILED due to: NA
2019-01-30 14:34:03,693 INFO [org.apache.hadoop.mapreduce.Job] - Counters: 0

Process finished with exit code 1

原因

在于自定义bean未实现Writable接口,导致map将output序列化到本地时调用write和read方法异常
本是低级错误,可见在学习过程中未能把握MapReduce工作流程、未能抓住基础。

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值