报错如下
(pyspark) cd12:~ cdarling$ pythonPython 3.7.3 (default, Mar 27 2019, 16:54:48)[Clang 4.0.1 (tags/RELEASE_401/final)] :: Anaconda, Inc. on darwinType "help", "copyright", "credits" or "license" for more information.>>> from pyspark.sql import SparkSession>>> s=SparkSession.builder.getOrCreate()19/07/13 14:05:59 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicableUsing Spark's default log4j profile: org/apache/spark/log4j-defaults.propertiesSetting default log level to "WARN".To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).19/07/13 14:06:00 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.19/07/13 14:06:00 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.19/07/13 14:06:00 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.19/07/13 14:06:00 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.19/07/13 14:06:00 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.19/07/13 14:06:00 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.19/07/13 14:06:00 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.19/07/13 14:06:00 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.19/07/13 14:06:00 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.19/07/13 14:06:00 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.19/07/13 14:06:00 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.19/07/13 14:06:00 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.19/07/13 14:06:00 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.19/07/13 14:06:00 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.19/07/13 14:06:00 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.19/07/13 14:06:00 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.19/07/13 14:06:00 ERROR SparkContext: Error initializing SparkContext.java.net.BindException: Can't assign requested address: Service 'sparkDriver' failed after 16 retries (on a random free port)! Consider explicitly setting the appropriate binding address for the service 'sparkDriver' (for example spark.driver.bindAddress for SparkDriver) to the correct binding address.at sun.nio.ch.Net.bind0(Native Method)at sun.nio.ch.Net.bind(Net.java:433)at sun.nio.ch.Net.bind(Net.java:425)at sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:223)at io.netty.channel.socket.nio.NioServerSocketChannel.doBind(NioServerSocketChannel.java:128)at io.netty.channel.AbstractChannel$AbstractUnsafe.bind(AbstractChannel.java:558)at io.netty.channel.DefaultChannelPipeline$HeadContext.bind(DefaultChannelPipeline.java:1283)at io.netty.channel.AbstractChannelHandlerContext.invokeBind(AbstractChannelHandlerContext.java:501)at io.netty.channel.AbstractChannelHandlerContext.bind(AbstractChannelHandlerContext.java:486)at io.netty.channel.DefaultChannelPipeline.bind(DefaultChannelPipeline.java:989)at io.netty.channel.AbstractChannel.bind(AbstractChannel.java:254)at io.netty.bootstrap.AbstractBootstrap$2.run(AbstractBootstrap.java:364)at io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:163)at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:403)at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:463)at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:858)at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:138)at java.lang.Thread.run(Thread.java:748)Traceback (most recent call last):File "<stdin>", line 1, in <module>File "/usr/local/Caskroom/miniconda/4.6.14/miniconda3/envs/pyspark/lib/python3.7/site-packages/pyspark/sql/session.py", line 173, in getOrCreatesc = SparkContext.getOrCreate(sparkConf)File "/usr/local/Caskroom/miniconda/4.6.14/miniconda3/envs/pyspark/lib/python3.7/site-packages/pyspark/context.py", line 367, in getOrCreateSparkContext(conf=conf or SparkConf())File "/usr/local/Caskroom/miniconda/4.6.14/miniconda3/envs/pyspark/lib/python3.7/site-packages/pyspark/context.py", line 136, in __init__conf, jsc, profiler_cls)File "/usr/local/Caskroom/miniconda/4.6.14/miniconda3/envs/pyspark/lib/python3.7/site-packages/pyspark/context.py", line 198, in _do_initself._jsc = jsc or self._initialize_context(self._conf._jconf)File "/usr/local/Caskroom/miniconda/4.6.14/miniconda3/envs/pyspark/lib/python3.7/site-packages/pyspark/context.py", line 306, in _initialize_contextreturn self._jvm.JavaSparkContext(jconf)File "/usr/local/Caskroom/miniconda/4.6.14/miniconda3/envs/pyspark/lib/python3.7/site-packages/py4j/java_gateway.py", line 1525, in __call__answer, self._gateway_client, None, self._fqn)File "/usr/local/Caskroom/miniconda/4.6.14/miniconda3/envs/pyspark/lib/python3.7/site-packages/py4j/protocol.py", line 328, in get_return_valueformat(target_id, ".", name), value)py4j.protocol.Py4JJavaError: An error occurred while calling None.org.apache.spark.api.java.JavaSparkContext.: java.net.BindException: Can't assign requested address: Service 'sparkDriver' failed after 16 retries (on a random free port)! Consider explicitly setting the appropriate binding address for the service 'sparkDriver' (for example spark.driver.bindAddress for SparkDriver) to the correct binding address.at sun.nio.ch.Net.bind0(Native Method)at sun.nio.ch.Net.bind(Net.java:433)at sun.nio.ch.Net.bind(Net.java:425)at sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:223)at io.netty.channel.socket.nio.NioServerSocketChannel.doBind(NioServerSocketChannel.java:128)at io.netty.channel.AbstractChannel$AbstractUnsafe.bind(AbstractChannel.java:558)at io.netty.channel.DefaultChannelPipeline$HeadContext.bind(DefaultChannelPipeline.java:1283)at io.netty.channel.AbstractChannelHandlerContext.invokeBind(AbstractChannelHandlerContext.java:501)at io.netty.channel.AbstractChannelHandlerContext.bind(AbstractChannelHandlerContext.java:486)at io.netty.channel.DefaultChannelPipeline.bind(DefaultChannelPipeline.java:989)at io.netty.channel.AbstractChannel.bind(AbstractChannel.java:254)at io.netty.bootstrap.AbstractBootstrap$2.run(AbstractBootstrap.java:364)at io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:163)at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:403)at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:463)at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:858)at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:138)at java.lang.Thread.run(Thread.java:748)>>>(pyspark) cd12:~ cdarling$
