Where communities thrive


  • Join over 1.5M+ people
  • Join over 100K+ communities
  • Free without limits
  • Create your own community
People
Activity
    xu20160924
    @xu20160924
    还是可以通过建一个MySQL的来兼容Doris表?
    luzhijing
    @luzhijing
    可以建成MySQL的外表
    Gitter使用得比较少,大家有问题可以在GitHub Discussion专区提问或者直接提交Issue
    Tgyhxy
    @Tgyhxy
    scala> mockDataDF1.write.format("doris"
    | ).option("doris.table.identifier", "xy.odbc_table3"
    | ).option("doris.fenodes", "192.168.0.152:8030"
    | ).option("user", "root"
    | ).option("password", "root123"
    | ).option("doris.write.fields","id"
    | ).save()
    java.lang.RuntimeException: org.apache.doris.spark.sql.DorisSourceProvider does not allow create table as select.
    at scala.sys.package$.error(package.scala:27)
    at org.apache.spark.sql.execution.datasources.DataSource.planForWriting(DataSource.scala:526)
    at org.apache.spark.sql.DataFrameWriter.saveToV1Source(DataFrameWriter.scala:281)
    at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:270)
    ... 57 elided
    大佬们好,我在Doris中建好表了,为什么一直写入写不进去呢?
    hilo
    @zhouhoo
    log4j漏洞问题,最近威胁很大,官方怎么还不发布修复版本呢
    1 reply
    log4j,我直接用2.15替换了2.14,fe报错跑不起来
    junweishiwo
    @junweishiwo
    各位好,有个问题请教大家:我最近在使用debezium结合routine load同步mysql数据到doris中,为了减少topic数量,我使用debezium将多个表的更新信息放到一个topic里了(json中有一个属性table,可以区分数据属于哪张表)。这种情况如何使用routine load将数据同步到doris中对应的表里呢,各位这样操作过吗?
    我看文档里说的都是一个topic一张表的情况
    CrazyDY
    @CrazyDY
    image.png
    2 replies
    请教一下各位大佬,在编译spark-connector的时候报这个错
    这个需要怎么解决呢
    CrazyDY
    @CrazyDY

    scala> mockDataDF1.write.format("doris"
    | ).option("doris.table.identifier", "xy.odbc_table3"
    | ).option("doris.fenodes", "192.168.0.152:8030"
    | ).option("user", "root"
    | ).option("password", "root123"
    | ).option("doris.write.fields","id"
    | ).save()
    java.lang.RuntimeException: org.apache.doris.spark.sql.DorisSourceProvider does not allow create table as select.
    at scala.sys.package$.error(package.scala:27)
    at org.apache.spark.sql.execution.datasources.DataSource.planForWriting(DataSource.scala:526)
    at org.apache.spark.sql.DataFrameWriter.saveToV1Source(DataFrameWriter.scala:281)
    at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:270)
    ... 57 elided
    大佬们好,我在Doris中建好表了,为什么一直写入写不进去呢?

    同问

    1 reply
    westgatepark
    @westgatepark
    有没有升级到doris 0.15 的大佬 稳定吗
    JiangJungle
    @JiangJungle
    doris缺少一个管理界面工具
    lin9long
    @lin9long
    doris 0.15 ,在已有的hadoop集群中,使用spark load,一直提示 configs/jobconfig.json文件不存在,观察etl执行过程这个文件是有生成的;yarn使用的是集群部署的客户端, 执行也一致报Error: JAVA_HOME is not set and could not be found.请问是什么原因导致的?
    1 reply
    panlijie
    @xiaoxiaopan118
    @JiangJungle 可以在ambari中管理,可以参考https://github.com/xiaoxiaopan118/Ambari-Doris-Service
    iMom0
    @imom0
    我有个疑惑,为什么官方提供了docker编译的镜像,不提供编译好的be和fe的镜像呢,是出于什么考虑
    sky
    @skymilong
    和楼上一样,自己编译很久
    luzhijing
    @luzhijing
    可以在palo.baidu.com 下载预编译好的palo发行版
    wanbaococ
    @winatox

    我有个疑惑,为什么官方提供了docker编译的镜像,不提供编译好的be和fe的镜像呢,是出于什么考虑

    因为编译和运行的环境指令集不同,我在自己的电脑上编译通过放到服务器上运行提示找不到指令,上网查了才知道两个的指令集不同,在服务器上再编译就通过了

    luzhijing
    @luzhijing
    Welcome to talk about some performance optimization tips of Apache Doris!
    luzhijing
    @luzhijing
    Click it and join the slack of Apache Doris Community!
    zhangwgch
    @zhangwgch
    streamLoad导入的时候,磁盘写入有限制吗,目前写入速度上不去,磁盘io10%都不到
    大神们,谁给解读一下
    zebingtian
    @zebingtian
    有大神在么
    Mingyu Chen
    @morningman

    streamLoad导入的时候,磁盘写入有限制吗,目前写入速度上不去,磁盘io10%都不到

    单个stream load是单线程的,如果数据量大,需要开多个并发导入

    青松
    @FireFreedomK

    各位大佬,萌新请教一个问题。我在三台机器上部署了doris,一台fe、be混部,两台单独部署be。doris版本是0.15.0

    在用spark DataFrame向doris导入数据时,一开始可以正常导入,但是后面会报如下的错误
    2022-01-26 14:29:41,867 ERROR Executor:94 Executor task launch worker for task 11714 718847 - Exception in task 9.0 in stage 8206.0 (TID 11714)
    org.apache.doris.spark.exception.StreamLoadException: stream load error: too many filtered rows
    at org.apache.doris.spark.DorisStreamLoad.load(DorisStreamLoad.java:162)
    at org.apache.doris.spark.DorisStreamLoad.load(DorisStreamLoad.java:149)
    at org.apache.doris.spark.sql.DorisSourceProvider

    KaTeX parse error: Can't use function '$' in math mode at position 8: anonfun$̲createRelation$…: anonfun$createRelation$1
    anonfun$org$apache$doris$spark$sql$DorisSourceProvideranonfunanonfunflush$1$1
    KaTeX parse error: Can't use function '$' in math mode at position 8: anonfun$̲apply$mcV$sp$1.…: anonfun$apply$mcV$sp$1.apply$mcVI$sp(DorisSourceProvider.scala:96)
            at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:160)
            at org.apache.doris.spark.sql.DorisSourceProvider
    anonfun$createRelation$1
    KaTeX parse error: Can't use function '$' in math mode at position 8: anonfun$̲org$apache$dori…: anonfun$org$apache$doris$spark$sql$DorisSourceProvider
    anonfun
    KaTeX parse error: Can't use function '$' in math mode at position 6: flush$̲1$1.apply$mcV$s…: flush$1$1.apply$mcV$sp(DorisSourceProvider.scala:86)
            at scala.util.control.Breaks.breakable(Breaks.scala:38)
            at org.apache.doris.spark.sql.DorisSourceProvider
    anonfun$createRelation$1.org$apache$doris$spark$sql$DorisSourceProvideranonfunanonfunflush$1(DorisSourceProvider.scala:84)
    at org.apache.doris.spark.sql.DorisSourceProvider
    KaTeX parse error: Can't use function '$' in math mode at position 8: anonfun$̲createRelation$…: anonfun$createRelation$1
    anonfun$apply$2.apply(DorisSourceProvider.scala:70)
    at org.apache.doris.spark.sql.DorisSourceProvider
    KaTeX parse error: Can't use function '$' in math mode at position 8: anonfun$̲createRelation$…: anonfun$createRelation$1
    anonfun$apply$2.apply(DorisSourceProvider.scala:62)
    at scala.collection.Iterator$class.foreach(Iterator.scala:743)
    at scala.collection.AbstractIterator.foreach(Iterator.scala:1174)
    at org.apache.doris.spark.sql.DorisSourceProvider
    KaTeX parse error: Can't use function '$' in math mode at position 8: anonfun$̲createRelation$…: anonfun$createRelation$1.apply(DorisSourceProvider.scala:62)
            at org.apache.doris.spark.sql.DorisSourceProvider
    anonfun$createRelation$1.apply(DorisSourceProvider.scala:60)
    at org.apache.spark.rdd.RDD
    KaTeX parse error: Can't use function '$' in math mode at position 8: anonfun$̲foreachPartitio…: anonfun$foreachPartition$1
    anonfun$apply$28.apply(RDD.scala:935)
    at org.apache.spark.rdd.RDD
    KaTeX parse error: Can't use function '$' in math mode at position 8: anonfun$̲foreachPartitio…: anonfun$foreachPartition$1
    anonfun$apply$28.apply(RDD.scala:935)
    at org.apache.spark.SparkContext
    KaTeX parse error: Can't use function '$' in math mode at position 8: anonfun$̲runJob$5.apply(…: anonfun$runJob$5.apply(SparkContext.scala:2121)
            at org.apache.spark.SparkContext
    anonfun$runJob$5.apply(SparkContext.scala:2121)
    at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
    at org.apache.spark.scheduler.Task.run(Task.scala:121)
    at org.apache.spark.executor.Executor$TaskRunner$$anonfun$11.apply(Executor.scala:407)
    at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1408)
    at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:413)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
    at java.lang.Thread.run(Thread.java:748)

    htyoung
    @htyoung
    image.png
    @FireFreedomK too many filtered rows说明导入的数据中有错误,被doris拒绝了,结合自己下面的错误日志排查一下原因
    jing zhang
    @kane0409_gitlab
    请问我搭建了三个节点,show backends;显示都alive,但有两个节点磁盘容量显示1b,这有问题吧?
    075a3e8eb78ebd61b89f9f00a56d1fd.png
    而且我点击最前面的BackendId,后两个都是没有信息,只有第一个有信息显示,能指导下大概是哪里的问题吗,多谢
    jing zhang
    @kane0409_gitlab
    在be的log里找到这个,后两个节点没找到master,这个在哪里配置呢?
    image.png
    luzhijing
    @luzhijing
    fe的priority_networks没生效
    青松
    @FireFreedomK
    @htyoung 好的,谢谢。
    青松
    @FireFreedomK
    @htyoung 我把那份数据下载下来看了一下,没有明显的异常。请问我在哪个位置找数据库的日志文件能帮助我定位问题?
    JiangJungle
    @JiangJungle
    @kane0409_gitlab 你用的啥界面工具?
    JiangJungle
    @JiangJungle
    今年上半年doris有哪些规划?
    JiangJungle
    @JiangJungle
    doris编程用什么?mysql驱动?像clickhouse有自己的驱动吗?
    Mingyu Chen
    @morningman
    The Doris Slack Channel is opened now. Welcome to join.
    The official Doris discussion channel will later be integrated into the Slack and dev@doris mailing lists. 后续Doris的官方讨论渠道将整合到 Slack 以及 dev@doris 邮件列表中。

    今年上半年doris有哪些规划?

    Roadmap 2022: apache/incubator-doris#7502

    AUB
    @aubdiy
    hdfs-broker 在导入 hdfs 数据的时候 失败, 日志中提示 找不到类
    java.lang.NoClassDefFoundError: Could not initialize class org.apache.hadoop.util.StringUtils
    有朋友遇到过吗
    doris 0.15, hadoop 3.3.1 hdfs-broker 在导入 hdfs 数据的时候 失败, 日志中提示 找不到类
    receive a ping request, request detail: TBrokerPingBrokerRequest(version:VERSION_ONE, clientId:172.31.3.146)
    [INFO ] 2022-02-08 08:07:10,947 method:org.apache.doris.broker.hdfs.HDFSBrokerServiceImpl.listPath(HDFSBrokerServiceImpl.java:67)
    received a list path request, request detail: TBrokerListPathRequest(version:VERSION_ONE, path:hdfs://test.internal:8020/tmp/data.csv, isRecursive:false, properties:{_DORIS_STORAGE_TYPE_=BROKER})
    [INFO ] 2022-02-08 08:07:10,948 method:org.apache.doris.broker.hdfs.FileSystemManager.getDistributedFileSystem(FileSystemManager.java:244)
    create file system for new path: hdfs://test.internal:8020/tmp/data.csv
    Exception in thread "pool-2-thread-13" java.lang.NoClassDefFoundError: Could not initialize class org.apache.hadoop.util.StringUtils
        at org.apache.hadoop.conf.Configuration.getBoolean(Configuration.java:1437)
        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:369)
        at org.apache.doris.broker.hdfs.FileSystemManager.getDistributedFileSystem(FileSystemManager.java:360)
        at org.apache.doris.broker.hdfs.FileSystemManager.getFileSystem(FileSystemManager.java:152)
        at org.apache.doris.broker.hdfs.FileSystemManager.listPath(FileSystemManager.java:427)
        at org.apache.doris.broker.hdfs.HDFSBrokerServiceImpl.listPath(HDFSBrokerServiceImpl.java:74)
        at org.apache.doris.thrift.TPaloBrokerService$Processor$listPath.getResult(TPaloBrokerService.java:815)
        at org.apache.doris.thrift.TPaloBrokerService$Processor$listPath.getResult(TPaloBrokerService.java:795)
        at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:38)
        at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:38)
        at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:313)
        at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
        at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
        at java.base/java.lang.Thread.run(Thread.java:834)
    [INFO ] 2022-02-08 08:07:10,963 method:org.apache.doris.broker.hdfs.HDFSBrokerServiceImpl.listPath(HDFSBrokerServiceImpl.java:67)
    received a list path request, request detail: TBrokerListPathRequest(version:VERSION_ONE, path:hdfs://test.internal:8020/tmp/data.csv, isRecursive:false, properties:{_DORIS_STORAGE_TYPE_=BROKER})